var/home/core/zuul-output/0000755000175000017500000000000015073135377014540 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015073145472015501 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004714372215073145463017717 0ustar rootrootOct 13 08:44:33 crc systemd[1]: Starting Kubernetes Kubelet... Oct 13 08:44:33 crc restorecon[4564]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 13 08:44:33 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 08:44:34 crc restorecon[4564]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 08:44:34 crc restorecon[4564]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 13 08:44:35 crc kubenswrapper[4685]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 13 08:44:35 crc kubenswrapper[4685]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 13 08:44:35 crc kubenswrapper[4685]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 13 08:44:35 crc kubenswrapper[4685]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 13 08:44:35 crc kubenswrapper[4685]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 13 08:44:35 crc kubenswrapper[4685]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.284568 4685 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287327 4685 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287342 4685 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287346 4685 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287350 4685 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287354 4685 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287358 4685 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287361 4685 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287372 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287376 4685 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287380 4685 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287383 4685 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287386 4685 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287390 4685 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287393 4685 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287397 4685 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287400 4685 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287403 4685 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287408 4685 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287412 4685 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287416 4685 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287419 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287423 4685 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287426 4685 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287429 4685 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287433 4685 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287436 4685 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287440 4685 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287443 4685 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287447 4685 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287450 4685 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287455 4685 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287462 4685 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287465 4685 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287469 4685 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287472 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287476 4685 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287479 4685 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287483 4685 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287486 4685 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287490 4685 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287494 4685 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287498 4685 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287502 4685 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287506 4685 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287510 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287514 4685 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287519 4685 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287523 4685 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287528 4685 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287532 4685 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287535 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287539 4685 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287542 4685 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287546 4685 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287550 4685 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287555 4685 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287558 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287563 4685 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287569 4685 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287572 4685 feature_gate.go:330] unrecognized feature gate: Example Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287577 4685 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287582 4685 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287586 4685 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287590 4685 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287596 4685 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287605 4685 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287614 4685 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287621 4685 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287625 4685 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287630 4685 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.287634 4685 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287772 4685 flags.go:64] FLAG: --address="0.0.0.0" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287786 4685 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287797 4685 flags.go:64] FLAG: --anonymous-auth="true" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287803 4685 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287809 4685 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287813 4685 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287819 4685 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287825 4685 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287829 4685 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287833 4685 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287838 4685 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287842 4685 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287847 4685 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287851 4685 flags.go:64] FLAG: --cgroup-root="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287855 4685 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287859 4685 flags.go:64] FLAG: --client-ca-file="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287864 4685 flags.go:64] FLAG: --cloud-config="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287868 4685 flags.go:64] FLAG: --cloud-provider="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287872 4685 flags.go:64] FLAG: --cluster-dns="[]" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287878 4685 flags.go:64] FLAG: --cluster-domain="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287882 4685 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287887 4685 flags.go:64] FLAG: --config-dir="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287891 4685 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287896 4685 flags.go:64] FLAG: --container-log-max-files="5" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287902 4685 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287907 4685 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287927 4685 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287933 4685 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287939 4685 flags.go:64] FLAG: --contention-profiling="false" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287944 4685 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287949 4685 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287955 4685 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287960 4685 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287965 4685 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287969 4685 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287973 4685 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287977 4685 flags.go:64] FLAG: --enable-load-reader="false" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287981 4685 flags.go:64] FLAG: --enable-server="true" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.287985 4685 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288034 4685 flags.go:64] FLAG: --event-burst="100" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288038 4685 flags.go:64] FLAG: --event-qps="50" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288043 4685 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288048 4685 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288052 4685 flags.go:64] FLAG: --eviction-hard="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288058 4685 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288062 4685 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288066 4685 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288070 4685 flags.go:64] FLAG: --eviction-soft="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288074 4685 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288078 4685 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288082 4685 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288087 4685 flags.go:64] FLAG: --experimental-mounter-path="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288091 4685 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288095 4685 flags.go:64] FLAG: --fail-swap-on="true" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288099 4685 flags.go:64] FLAG: --feature-gates="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288104 4685 flags.go:64] FLAG: --file-check-frequency="20s" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288108 4685 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288113 4685 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288118 4685 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288122 4685 flags.go:64] FLAG: --healthz-port="10248" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288126 4685 flags.go:64] FLAG: --help="false" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288130 4685 flags.go:64] FLAG: --hostname-override="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288134 4685 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288139 4685 flags.go:64] FLAG: --http-check-frequency="20s" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288143 4685 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288147 4685 flags.go:64] FLAG: --image-credential-provider-config="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288151 4685 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288154 4685 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288159 4685 flags.go:64] FLAG: --image-service-endpoint="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288164 4685 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288168 4685 flags.go:64] FLAG: --kube-api-burst="100" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288172 4685 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288176 4685 flags.go:64] FLAG: --kube-api-qps="50" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288180 4685 flags.go:64] FLAG: --kube-reserved="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288184 4685 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288188 4685 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288192 4685 flags.go:64] FLAG: --kubelet-cgroups="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288196 4685 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288200 4685 flags.go:64] FLAG: --lock-file="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288204 4685 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288208 4685 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288212 4685 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288218 4685 flags.go:64] FLAG: --log-json-split-stream="false" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288223 4685 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288227 4685 flags.go:64] FLAG: --log-text-split-stream="false" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288231 4685 flags.go:64] FLAG: --logging-format="text" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288236 4685 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288240 4685 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288244 4685 flags.go:64] FLAG: --manifest-url="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288249 4685 flags.go:64] FLAG: --manifest-url-header="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288255 4685 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288258 4685 flags.go:64] FLAG: --max-open-files="1000000" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288264 4685 flags.go:64] FLAG: --max-pods="110" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288268 4685 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288272 4685 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288276 4685 flags.go:64] FLAG: --memory-manager-policy="None" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288280 4685 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288284 4685 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288288 4685 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288292 4685 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288302 4685 flags.go:64] FLAG: --node-status-max-images="50" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288306 4685 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288311 4685 flags.go:64] FLAG: --oom-score-adj="-999" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288315 4685 flags.go:64] FLAG: --pod-cidr="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288319 4685 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288326 4685 flags.go:64] FLAG: --pod-manifest-path="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288330 4685 flags.go:64] FLAG: --pod-max-pids="-1" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288335 4685 flags.go:64] FLAG: --pods-per-core="0" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288338 4685 flags.go:64] FLAG: --port="10250" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288342 4685 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288347 4685 flags.go:64] FLAG: --provider-id="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288351 4685 flags.go:64] FLAG: --qos-reserved="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288355 4685 flags.go:64] FLAG: --read-only-port="10255" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288359 4685 flags.go:64] FLAG: --register-node="true" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288363 4685 flags.go:64] FLAG: --register-schedulable="true" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288367 4685 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288375 4685 flags.go:64] FLAG: --registry-burst="10" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288379 4685 flags.go:64] FLAG: --registry-qps="5" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288383 4685 flags.go:64] FLAG: --reserved-cpus="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288387 4685 flags.go:64] FLAG: --reserved-memory="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288393 4685 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288397 4685 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288401 4685 flags.go:64] FLAG: --rotate-certificates="false" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288405 4685 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288409 4685 flags.go:64] FLAG: --runonce="false" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288413 4685 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288417 4685 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288422 4685 flags.go:64] FLAG: --seccomp-default="false" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288425 4685 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288429 4685 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288433 4685 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288438 4685 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288442 4685 flags.go:64] FLAG: --storage-driver-password="root" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288446 4685 flags.go:64] FLAG: --storage-driver-secure="false" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288450 4685 flags.go:64] FLAG: --storage-driver-table="stats" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288454 4685 flags.go:64] FLAG: --storage-driver-user="root" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288459 4685 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288463 4685 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288467 4685 flags.go:64] FLAG: --system-cgroups="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288471 4685 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288478 4685 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288482 4685 flags.go:64] FLAG: --tls-cert-file="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288486 4685 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288491 4685 flags.go:64] FLAG: --tls-min-version="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288495 4685 flags.go:64] FLAG: --tls-private-key-file="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288499 4685 flags.go:64] FLAG: --topology-manager-policy="none" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288503 4685 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288507 4685 flags.go:64] FLAG: --topology-manager-scope="container" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288511 4685 flags.go:64] FLAG: --v="2" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288517 4685 flags.go:64] FLAG: --version="false" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288522 4685 flags.go:64] FLAG: --vmodule="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288527 4685 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288531 4685 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288631 4685 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288636 4685 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288640 4685 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288643 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288647 4685 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288650 4685 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288654 4685 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288658 4685 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288661 4685 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288665 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288668 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288672 4685 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288675 4685 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288678 4685 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288682 4685 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288685 4685 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288689 4685 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288692 4685 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288695 4685 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288699 4685 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288703 4685 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288706 4685 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288715 4685 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288719 4685 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288723 4685 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288727 4685 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288730 4685 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288734 4685 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288738 4685 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288742 4685 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288746 4685 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288750 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288754 4685 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288759 4685 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288763 4685 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288767 4685 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288770 4685 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288775 4685 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288779 4685 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288784 4685 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288789 4685 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288793 4685 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288796 4685 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288800 4685 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288804 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288807 4685 feature_gate.go:330] unrecognized feature gate: Example Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288811 4685 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288815 4685 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288818 4685 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288822 4685 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288825 4685 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288829 4685 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288832 4685 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288836 4685 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288842 4685 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288846 4685 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288850 4685 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288853 4685 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288856 4685 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288861 4685 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288864 4685 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288868 4685 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288871 4685 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288875 4685 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288882 4685 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288886 4685 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288890 4685 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288893 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288896 4685 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288901 4685 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.288905 4685 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.288937 4685 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.296393 4685 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.296429 4685 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296496 4685 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296504 4685 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296508 4685 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296513 4685 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296518 4685 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296522 4685 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296525 4685 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296529 4685 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296534 4685 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296541 4685 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296546 4685 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296550 4685 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296554 4685 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296558 4685 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296562 4685 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296565 4685 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296569 4685 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296572 4685 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296576 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296579 4685 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296583 4685 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296588 4685 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296593 4685 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296598 4685 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296602 4685 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296606 4685 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296610 4685 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296614 4685 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296617 4685 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296621 4685 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296624 4685 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296628 4685 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296631 4685 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296635 4685 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296640 4685 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296643 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296647 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296650 4685 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296654 4685 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296657 4685 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296661 4685 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296664 4685 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296668 4685 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296671 4685 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296675 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296679 4685 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296682 4685 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296686 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296690 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296694 4685 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296697 4685 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296702 4685 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296706 4685 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296710 4685 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296714 4685 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296720 4685 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296724 4685 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296728 4685 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296732 4685 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296736 4685 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296741 4685 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296745 4685 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296749 4685 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296752 4685 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296757 4685 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296762 4685 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296766 4685 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296771 4685 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296775 4685 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296778 4685 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296783 4685 feature_gate.go:330] unrecognized feature gate: Example Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.296790 4685 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296896 4685 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296903 4685 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296911 4685 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296929 4685 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296932 4685 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296936 4685 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296940 4685 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296943 4685 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296947 4685 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296951 4685 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296955 4685 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296959 4685 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296962 4685 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296966 4685 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296969 4685 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296974 4685 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296978 4685 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296982 4685 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296986 4685 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296991 4685 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296995 4685 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.296999 4685 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297003 4685 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297007 4685 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297011 4685 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297015 4685 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297020 4685 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297024 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297028 4685 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297033 4685 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297038 4685 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297042 4685 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297047 4685 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297051 4685 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297059 4685 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297063 4685 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297068 4685 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297072 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297076 4685 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297080 4685 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297083 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297087 4685 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297091 4685 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297094 4685 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297098 4685 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297102 4685 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297105 4685 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297109 4685 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297113 4685 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297117 4685 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297120 4685 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297124 4685 feature_gate.go:330] unrecognized feature gate: Example Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297128 4685 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297131 4685 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297135 4685 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297139 4685 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297142 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297145 4685 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297149 4685 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297152 4685 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297156 4685 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297159 4685 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297163 4685 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297166 4685 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297170 4685 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297174 4685 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297177 4685 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297180 4685 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297184 4685 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297187 4685 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.297191 4685 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.297198 4685 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.298832 4685 server.go:940] "Client rotation is on, will bootstrap in background" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.302848 4685 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.302958 4685 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.305103 4685 server.go:997] "Starting client certificate rotation" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.305128 4685 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.305344 4685 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-03 00:54:23.514424632 +0000 UTC Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.305502 4685 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1216h9m48.208926645s for next certificate rotation Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.333436 4685 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.338672 4685 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.352875 4685 log.go:25] "Validated CRI v1 runtime API" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.385477 4685 log.go:25] "Validated CRI v1 image API" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.386790 4685 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.392775 4685 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-13-08-39-27-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.392801 4685 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.406248 4685 manager.go:217] Machine: {Timestamp:2025-10-13 08:44:35.402532651 +0000 UTC m=+0.550408432 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2799998 MemoryCapacity:25199480832 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:1ab45c59-0bc0-4269-a6a9-8fdefb164798 BootID:774b74cd-e138-4e5c-928f-324075caa6f2 Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599742464 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:77:82:dc Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:77:82:dc Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:14:03:df Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:11:e4:07 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:1f:db:a4 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:26:74:11 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:22:07:9b:88:01:3f Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:36:6e:28:10:be:29 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199480832 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.406428 4685 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.406575 4685 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.409044 4685 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.409252 4685 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.409297 4685 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.409484 4685 topology_manager.go:138] "Creating topology manager with none policy" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.409495 4685 container_manager_linux.go:303] "Creating device plugin manager" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.410199 4685 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.410229 4685 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.411082 4685 state_mem.go:36] "Initialized new in-memory state store" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.411178 4685 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.415086 4685 kubelet.go:418] "Attempting to sync node with API server" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.415107 4685 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.415125 4685 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.415139 4685 kubelet.go:324] "Adding apiserver pod source" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.415148 4685 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.424861 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.424905 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Oct 13 08:44:35 crc kubenswrapper[4685]: E1013 08:44:35.425022 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.65:6443: connect: connection refused" logger="UnhandledError" Oct 13 08:44:35 crc kubenswrapper[4685]: E1013 08:44:35.425097 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.65:6443: connect: connection refused" logger="UnhandledError" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.425228 4685 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.426447 4685 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.428096 4685 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.430304 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.430331 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.430340 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.430351 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.430365 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.430372 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.430381 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.430396 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.430406 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.430415 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.430439 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.430447 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.430475 4685 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.430965 4685 server.go:1280] "Started kubelet" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.431669 4685 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.432766 4685 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.432998 4685 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 13 08:44:35 crc systemd[1]: Started Kubernetes Kubelet. Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.433589 4685 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.436642 4685 server.go:460] "Adding debug handlers to kubelet server" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.440950 4685 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.441103 4685 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.441255 4685 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 13:57:15.734267567 +0000 UTC Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.441294 4685 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1493h12m40.292975589s for next certificate rotation Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.441708 4685 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.441747 4685 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.441905 4685 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 13 08:44:35 crc kubenswrapper[4685]: E1013 08:44:35.439717 4685 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.65:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186e0084f72b5567 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-13 08:44:35.430937959 +0000 UTC m=+0.578813730,LastTimestamp:2025-10-13 08:44:35.430937959 +0000 UTC m=+0.578813730,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 13 08:44:35 crc kubenswrapper[4685]: E1013 08:44:35.442466 4685 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.443169 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Oct 13 08:44:35 crc kubenswrapper[4685]: E1013 08:44:35.444099 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.65:6443: connect: connection refused" logger="UnhandledError" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.443843 4685 factory.go:55] Registering systemd factory Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.444302 4685 factory.go:221] Registration of the systemd container factory successfully Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.444626 4685 factory.go:153] Registering CRI-O factory Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.445392 4685 factory.go:221] Registration of the crio container factory successfully Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.445508 4685 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.445590 4685 factory.go:103] Registering Raw factory Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.445647 4685 manager.go:1196] Started watching for new ooms in manager Oct 13 08:44:35 crc kubenswrapper[4685]: E1013 08:44:35.445054 4685 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.65:6443: connect: connection refused" interval="200ms" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.446341 4685 manager.go:319] Starting recovery of all containers Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.471701 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.471780 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.471793 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.471802 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.471828 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.471838 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.471847 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.471857 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.471867 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.471877 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.471886 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.471918 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.471928 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.471938 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.471948 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.471957 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.471981 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.471990 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.471999 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472008 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472018 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472028 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472037 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472061 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472071 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472080 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472091 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472101 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472111 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472148 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472157 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472166 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472179 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472187 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472227 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472236 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472247 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472256 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472267 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472276 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472317 4685 manager.go:324] Recovery completed Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472326 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472455 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472484 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472498 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472512 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472526 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472538 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472554 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472569 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472581 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472594 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472607 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472629 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472643 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472659 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472672 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472686 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472699 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472712 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472723 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472738 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472753 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472765 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472780 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472793 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472807 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472822 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472836 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472850 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472866 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472882 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472895 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472910 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472945 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472961 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472973 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.472988 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473001 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473014 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473030 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473044 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473086 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473101 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473115 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473130 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473143 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473158 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473207 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473223 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473237 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473250 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473264 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473278 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473291 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473306 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473322 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473336 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473349 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473363 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473376 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473390 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473404 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473419 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473433 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473456 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473469 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473482 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473496 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473509 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473523 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473538 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473552 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473566 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473580 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473594 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473608 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473619 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473633 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473647 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473660 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473674 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473688 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473702 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473715 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473730 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473745 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473759 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473775 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473788 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473802 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473818 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473831 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473846 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473860 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473876 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473890 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473903 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473934 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473949 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473963 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473976 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.473989 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474002 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474016 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474032 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474045 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474058 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474071 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474085 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474098 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474114 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474127 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474139 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474153 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474166 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474180 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474193 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474204 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474218 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474230 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474244 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474257 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474271 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474283 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474296 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474308 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474321 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474332 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474345 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474358 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474371 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474385 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474398 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.474411 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477091 4685 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477126 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477162 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477176 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477189 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477206 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477238 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477252 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477267 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477279 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477292 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477325 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477338 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477352 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477365 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477395 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477408 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477422 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477433 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477445 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477478 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477503 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477516 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477527 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477558 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477569 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477580 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477592 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477603 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477637 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477649 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477663 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477674 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477688 4685 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477716 4685 reconstruct.go:97] "Volume reconstruction finished" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.477725 4685 reconciler.go:26] "Reconciler: start to sync state" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.483188 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.487227 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.487364 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.487427 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.488160 4685 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.488178 4685 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.488199 4685 state_mem.go:36] "Initialized new in-memory state store" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.498646 4685 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.501393 4685 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.501425 4685 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.501452 4685 kubelet.go:2335] "Starting kubelet main sync loop" Oct 13 08:44:35 crc kubenswrapper[4685]: E1013 08:44:35.501489 4685 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.502436 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Oct 13 08:44:35 crc kubenswrapper[4685]: E1013 08:44:35.502487 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.65:6443: connect: connection refused" logger="UnhandledError" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.506839 4685 policy_none.go:49] "None policy: Start" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.507707 4685 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.507733 4685 state_mem.go:35] "Initializing new in-memory state store" Oct 13 08:44:35 crc kubenswrapper[4685]: E1013 08:44:35.542899 4685 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.573370 4685 manager.go:334] "Starting Device Plugin manager" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.573432 4685 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.573447 4685 server.go:79] "Starting device plugin registration server" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.573940 4685 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.573959 4685 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.574159 4685 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.574241 4685 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.574250 4685 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 13 08:44:35 crc kubenswrapper[4685]: E1013 08:44:35.581673 4685 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.602036 4685 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.602125 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.603713 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.603757 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.603771 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.603949 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.604528 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.604596 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.604769 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.604789 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.604796 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.605153 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.605697 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.605755 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.606275 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.606314 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.606327 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.606438 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.606464 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.606479 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.606524 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.606824 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.606843 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.606856 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.607124 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.607151 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.607604 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.607637 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.607659 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.607784 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.608244 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.608280 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.608731 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.608759 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.608769 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.608974 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.608994 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.609003 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.609122 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.609175 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.609187 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.609197 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.609203 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.610378 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.610402 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.610414 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:35 crc kubenswrapper[4685]: E1013 08:44:35.646850 4685 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.65:6443: connect: connection refused" interval="400ms" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.674277 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.675763 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.675812 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.675826 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.675860 4685 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 13 08:44:35 crc kubenswrapper[4685]: E1013 08:44:35.676538 4685 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.65:6443: connect: connection refused" node="crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.681934 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.682076 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.682190 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.682290 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.682368 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.682468 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.682602 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.682708 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.682803 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.682889 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.683208 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.683337 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.683436 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.683561 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.683710 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.785156 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.785218 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.785241 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.785257 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.785272 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.785313 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.785335 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.785356 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.785443 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.785465 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.785485 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.785526 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.785548 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.785569 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.785589 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.786051 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.786094 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.786111 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.786131 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.786147 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.786165 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.786184 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.786207 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.786244 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.786269 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.786300 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.786324 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.786352 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.786385 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.786409 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.876852 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.878582 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.878691 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.878752 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.878819 4685 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 13 08:44:35 crc kubenswrapper[4685]: E1013 08:44:35.879265 4685 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.65:6443: connect: connection refused" node="crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.931552 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.946278 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.961194 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.967378 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: I1013 08:44:35.971728 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.983450 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-74c9ab884a57e3e7aeca739810c35d8897906141fec0c2b541133de3c1ac027f WatchSource:0}: Error finding container 74c9ab884a57e3e7aeca739810c35d8897906141fec0c2b541133de3c1ac027f: Status 404 returned error can't find the container with id 74c9ab884a57e3e7aeca739810c35d8897906141fec0c2b541133de3c1ac027f Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.987171 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-4ac10ac39c71753ff0552cd371636fef6929b1ec2034c4621e27aa547e1b74d5 WatchSource:0}: Error finding container 4ac10ac39c71753ff0552cd371636fef6929b1ec2034c4621e27aa547e1b74d5: Status 404 returned error can't find the container with id 4ac10ac39c71753ff0552cd371636fef6929b1ec2034c4621e27aa547e1b74d5 Oct 13 08:44:35 crc kubenswrapper[4685]: W1013 08:44:35.996393 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-ea3fb5d09a5e26449d36e86ab3208c88f0f6e32c2671164636bb6dbae4cf6b0d WatchSource:0}: Error finding container ea3fb5d09a5e26449d36e86ab3208c88f0f6e32c2671164636bb6dbae4cf6b0d: Status 404 returned error can't find the container with id ea3fb5d09a5e26449d36e86ab3208c88f0f6e32c2671164636bb6dbae4cf6b0d Oct 13 08:44:36 crc kubenswrapper[4685]: W1013 08:44:36.003124 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-b369c10b42d3b5aaf657b6952c8fe4f50d26c1b8f93fee77f2d7e63c6ab5b859 WatchSource:0}: Error finding container b369c10b42d3b5aaf657b6952c8fe4f50d26c1b8f93fee77f2d7e63c6ab5b859: Status 404 returned error can't find the container with id b369c10b42d3b5aaf657b6952c8fe4f50d26c1b8f93fee77f2d7e63c6ab5b859 Oct 13 08:44:36 crc kubenswrapper[4685]: E1013 08:44:36.048077 4685 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.65:6443: connect: connection refused" interval="800ms" Oct 13 08:44:36 crc kubenswrapper[4685]: I1013 08:44:36.279604 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:36 crc kubenswrapper[4685]: I1013 08:44:36.280770 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:36 crc kubenswrapper[4685]: I1013 08:44:36.280822 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:36 crc kubenswrapper[4685]: I1013 08:44:36.280835 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:36 crc kubenswrapper[4685]: I1013 08:44:36.280866 4685 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 13 08:44:36 crc kubenswrapper[4685]: E1013 08:44:36.281383 4685 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.65:6443: connect: connection refused" node="crc" Oct 13 08:44:36 crc kubenswrapper[4685]: I1013 08:44:36.433005 4685 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Oct 13 08:44:36 crc kubenswrapper[4685]: W1013 08:44:36.500371 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Oct 13 08:44:36 crc kubenswrapper[4685]: E1013 08:44:36.500474 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.65:6443: connect: connection refused" logger="UnhandledError" Oct 13 08:44:36 crc kubenswrapper[4685]: I1013 08:44:36.505169 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ea3fb5d09a5e26449d36e86ab3208c88f0f6e32c2671164636bb6dbae4cf6b0d"} Oct 13 08:44:36 crc kubenswrapper[4685]: I1013 08:44:36.507056 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"74c9ab884a57e3e7aeca739810c35d8897906141fec0c2b541133de3c1ac027f"} Oct 13 08:44:36 crc kubenswrapper[4685]: I1013 08:44:36.508452 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"4ac10ac39c71753ff0552cd371636fef6929b1ec2034c4621e27aa547e1b74d5"} Oct 13 08:44:36 crc kubenswrapper[4685]: I1013 08:44:36.510724 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"dee6003be9656885d7ca5a8e0eabddeeeef246f238b1e0ebda35afa2bad60722"} Oct 13 08:44:36 crc kubenswrapper[4685]: I1013 08:44:36.512849 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b369c10b42d3b5aaf657b6952c8fe4f50d26c1b8f93fee77f2d7e63c6ab5b859"} Oct 13 08:44:36 crc kubenswrapper[4685]: W1013 08:44:36.713016 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Oct 13 08:44:36 crc kubenswrapper[4685]: E1013 08:44:36.713352 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.65:6443: connect: connection refused" logger="UnhandledError" Oct 13 08:44:36 crc kubenswrapper[4685]: W1013 08:44:36.738561 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Oct 13 08:44:36 crc kubenswrapper[4685]: E1013 08:44:36.738652 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.65:6443: connect: connection refused" logger="UnhandledError" Oct 13 08:44:36 crc kubenswrapper[4685]: E1013 08:44:36.849012 4685 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.65:6443: connect: connection refused" interval="1.6s" Oct 13 08:44:37 crc kubenswrapper[4685]: W1013 08:44:37.011052 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Oct 13 08:44:37 crc kubenswrapper[4685]: E1013 08:44:37.011149 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.65:6443: connect: connection refused" logger="UnhandledError" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.082174 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.084110 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.084155 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.084171 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.084206 4685 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 13 08:44:37 crc kubenswrapper[4685]: E1013 08:44:37.084734 4685 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.65:6443: connect: connection refused" node="crc" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.433279 4685 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.516745 4685 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488" exitCode=0 Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.516834 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.516843 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488"} Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.517987 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.518013 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.518022 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.519985 4685 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="8638dfc388b790dc983b512805c3d49c7c25b04ec7708d7d6eb92fd324a6a123" exitCode=0 Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.520112 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"8638dfc388b790dc983b512805c3d49c7c25b04ec7708d7d6eb92fd324a6a123"} Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.520306 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.521317 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.521447 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.521506 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.523165 4685 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59" exitCode=0 Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.523284 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59"} Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.523419 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.524233 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.524318 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.524378 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.527181 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525"} Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.527306 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0"} Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.527397 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367"} Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.527476 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca"} Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.527625 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.529334 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.529358 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.529370 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.531028 4685 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee" exitCode=0 Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.531063 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee"} Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.531115 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.531726 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.531751 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.531760 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.547563 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.548674 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.548730 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:37 crc kubenswrapper[4685]: I1013 08:44:37.548743 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.432483 4685 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Oct 13 08:44:38 crc kubenswrapper[4685]: E1013 08:44:38.450703 4685 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.65:6443: connect: connection refused" interval="3.2s" Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.541211 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d"} Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.541314 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6"} Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.541349 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e"} Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.543947 4685 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275" exitCode=0 Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.544095 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275"} Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.544250 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.545391 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.545431 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.545442 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.548888 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"39c698197c289f5a88b55f58360dde69af247a5a1697552450f462c47019be83"} Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.549298 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.553218 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.553266 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.553280 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.563343 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"16b91dfe5b7be866dc336876627a48fb7f5c3d4cfca022c4f7cf2cd448bb9510"} Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.563416 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.563435 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.563413 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"305186b83e0255ae886edebdb2038898352c016ab17ed4a594c7caac64babde0"} Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.563803 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e7d110b9d78594b2f634d07714a305066c53a277b10a9d2ef513e47fc19f52b3"} Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.564456 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.564460 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.564489 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.564507 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.564516 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.564519 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.684994 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.686417 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.686453 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.686466 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:38 crc kubenswrapper[4685]: I1013 08:44:38.686494 4685 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 13 08:44:38 crc kubenswrapper[4685]: E1013 08:44:38.686945 4685 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.65:6443: connect: connection refused" node="crc" Oct 13 08:44:38 crc kubenswrapper[4685]: W1013 08:44:38.701959 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Oct 13 08:44:38 crc kubenswrapper[4685]: E1013 08:44:38.702033 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.65:6443: connect: connection refused" logger="UnhandledError" Oct 13 08:44:39 crc kubenswrapper[4685]: W1013 08:44:39.178109 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Oct 13 08:44:39 crc kubenswrapper[4685]: E1013 08:44:39.178289 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.65:6443: connect: connection refused" logger="UnhandledError" Oct 13 08:44:39 crc kubenswrapper[4685]: W1013 08:44:39.201432 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Oct 13 08:44:39 crc kubenswrapper[4685]: E1013 08:44:39.201568 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.65:6443: connect: connection refused" logger="UnhandledError" Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.568010 4685 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4" exitCode=0 Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.568620 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4"} Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.568819 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.569826 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.569878 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.569897 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.574828 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.575129 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f"} Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.575181 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8"} Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.577351 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.578330 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.578392 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.582888 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.583083 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.583178 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.583361 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.583417 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.583433 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.584298 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.584409 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.584503 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.802876 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.803065 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.804109 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.804161 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:39 crc kubenswrapper[4685]: I1013 08:44:39.804175 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:40 crc kubenswrapper[4685]: I1013 08:44:40.348209 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 08:44:40 crc kubenswrapper[4685]: I1013 08:44:40.355418 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 08:44:40 crc kubenswrapper[4685]: I1013 08:44:40.582125 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 08:44:40 crc kubenswrapper[4685]: I1013 08:44:40.582173 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:40 crc kubenswrapper[4685]: I1013 08:44:40.582520 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:40 crc kubenswrapper[4685]: I1013 08:44:40.582768 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed"} Oct 13 08:44:40 crc kubenswrapper[4685]: I1013 08:44:40.582839 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11"} Oct 13 08:44:40 crc kubenswrapper[4685]: I1013 08:44:40.582851 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8"} Oct 13 08:44:40 crc kubenswrapper[4685]: I1013 08:44:40.582860 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f"} Oct 13 08:44:40 crc kubenswrapper[4685]: I1013 08:44:40.582867 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4"} Oct 13 08:44:40 crc kubenswrapper[4685]: I1013 08:44:40.582900 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:40 crc kubenswrapper[4685]: I1013 08:44:40.582951 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 08:44:40 crc kubenswrapper[4685]: I1013 08:44:40.583488 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:40 crc kubenswrapper[4685]: I1013 08:44:40.583579 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:40 crc kubenswrapper[4685]: I1013 08:44:40.583584 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:40 crc kubenswrapper[4685]: I1013 08:44:40.583596 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:40 crc kubenswrapper[4685]: I1013 08:44:40.583613 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:40 crc kubenswrapper[4685]: I1013 08:44:40.583604 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:40 crc kubenswrapper[4685]: I1013 08:44:40.584262 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:40 crc kubenswrapper[4685]: I1013 08:44:40.584285 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:40 crc kubenswrapper[4685]: I1013 08:44:40.584295 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:41 crc kubenswrapper[4685]: I1013 08:44:41.229501 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 08:44:41 crc kubenswrapper[4685]: I1013 08:44:41.584863 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:41 crc kubenswrapper[4685]: I1013 08:44:41.584880 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:41 crc kubenswrapper[4685]: I1013 08:44:41.585059 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:41 crc kubenswrapper[4685]: I1013 08:44:41.585830 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:41 crc kubenswrapper[4685]: I1013 08:44:41.585856 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:41 crc kubenswrapper[4685]: I1013 08:44:41.585864 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:41 crc kubenswrapper[4685]: I1013 08:44:41.586260 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:41 crc kubenswrapper[4685]: I1013 08:44:41.586316 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:41 crc kubenswrapper[4685]: I1013 08:44:41.586341 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:41 crc kubenswrapper[4685]: I1013 08:44:41.586702 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:41 crc kubenswrapper[4685]: I1013 08:44:41.586720 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:41 crc kubenswrapper[4685]: I1013 08:44:41.586728 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:41 crc kubenswrapper[4685]: I1013 08:44:41.887904 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:41 crc kubenswrapper[4685]: I1013 08:44:41.889633 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:41 crc kubenswrapper[4685]: I1013 08:44:41.889693 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:41 crc kubenswrapper[4685]: I1013 08:44:41.889706 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:41 crc kubenswrapper[4685]: I1013 08:44:41.889747 4685 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 13 08:44:41 crc kubenswrapper[4685]: I1013 08:44:41.997809 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 13 08:44:42 crc kubenswrapper[4685]: I1013 08:44:42.043712 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 08:44:42 crc kubenswrapper[4685]: I1013 08:44:42.588316 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:42 crc kubenswrapper[4685]: I1013 08:44:42.589271 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:42 crc kubenswrapper[4685]: I1013 08:44:42.590479 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:42 crc kubenswrapper[4685]: I1013 08:44:42.590532 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:42 crc kubenswrapper[4685]: I1013 08:44:42.590549 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:42 crc kubenswrapper[4685]: I1013 08:44:42.591261 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:42 crc kubenswrapper[4685]: I1013 08:44:42.591382 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:42 crc kubenswrapper[4685]: I1013 08:44:42.591414 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:42 crc kubenswrapper[4685]: I1013 08:44:42.798439 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 08:44:42 crc kubenswrapper[4685]: I1013 08:44:42.798773 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:42 crc kubenswrapper[4685]: I1013 08:44:42.800788 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:42 crc kubenswrapper[4685]: I1013 08:44:42.800885 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:42 crc kubenswrapper[4685]: I1013 08:44:42.800954 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:42 crc kubenswrapper[4685]: I1013 08:44:42.806130 4685 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 13 08:44:42 crc kubenswrapper[4685]: I1013 08:44:42.806280 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 13 08:44:43 crc kubenswrapper[4685]: I1013 08:44:43.046595 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 08:44:43 crc kubenswrapper[4685]: I1013 08:44:43.589775 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:43 crc kubenswrapper[4685]: I1013 08:44:43.591318 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:43 crc kubenswrapper[4685]: I1013 08:44:43.591388 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:43 crc kubenswrapper[4685]: I1013 08:44:43.591420 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:44 crc kubenswrapper[4685]: I1013 08:44:44.429955 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 13 08:44:44 crc kubenswrapper[4685]: I1013 08:44:44.430139 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:44 crc kubenswrapper[4685]: I1013 08:44:44.431237 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:44 crc kubenswrapper[4685]: I1013 08:44:44.431269 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:44 crc kubenswrapper[4685]: I1013 08:44:44.431281 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:45 crc kubenswrapper[4685]: I1013 08:44:45.014263 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 13 08:44:45 crc kubenswrapper[4685]: I1013 08:44:45.014563 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:45 crc kubenswrapper[4685]: I1013 08:44:45.016548 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:45 crc kubenswrapper[4685]: I1013 08:44:45.016604 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:45 crc kubenswrapper[4685]: I1013 08:44:45.016622 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:45 crc kubenswrapper[4685]: E1013 08:44:45.581808 4685 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 13 08:44:47 crc kubenswrapper[4685]: I1013 08:44:47.351137 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 08:44:47 crc kubenswrapper[4685]: I1013 08:44:47.351273 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:47 crc kubenswrapper[4685]: I1013 08:44:47.352510 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:47 crc kubenswrapper[4685]: I1013 08:44:47.352571 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:47 crc kubenswrapper[4685]: I1013 08:44:47.352585 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:49 crc kubenswrapper[4685]: W1013 08:44:49.404619 4685 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 13 08:44:49 crc kubenswrapper[4685]: I1013 08:44:49.404756 4685 trace.go:236] Trace[1801490]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (13-Oct-2025 08:44:39.403) (total time: 10001ms): Oct 13 08:44:49 crc kubenswrapper[4685]: Trace[1801490]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (08:44:49.404) Oct 13 08:44:49 crc kubenswrapper[4685]: Trace[1801490]: [10.001527587s] [10.001527587s] END Oct 13 08:44:49 crc kubenswrapper[4685]: E1013 08:44:49.404792 4685 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 13 08:44:49 crc kubenswrapper[4685]: I1013 08:44:49.434062 4685 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 13 08:44:49 crc kubenswrapper[4685]: I1013 08:44:49.506491 4685 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 13 08:44:49 crc kubenswrapper[4685]: I1013 08:44:49.506553 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 13 08:44:49 crc kubenswrapper[4685]: I1013 08:44:49.511014 4685 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 13 08:44:49 crc kubenswrapper[4685]: I1013 08:44:49.511086 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 13 08:44:52 crc kubenswrapper[4685]: I1013 08:44:52.803676 4685 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 13 08:44:52 crc kubenswrapper[4685]: I1013 08:44:52.803799 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 13 08:44:52 crc kubenswrapper[4685]: I1013 08:44:52.811833 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 08:44:52 crc kubenswrapper[4685]: I1013 08:44:52.812042 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:52 crc kubenswrapper[4685]: I1013 08:44:52.813334 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:52 crc kubenswrapper[4685]: I1013 08:44:52.813414 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:52 crc kubenswrapper[4685]: I1013 08:44:52.813436 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:52 crc kubenswrapper[4685]: I1013 08:44:52.819176 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 08:44:53 crc kubenswrapper[4685]: I1013 08:44:53.618374 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 08:44:53 crc kubenswrapper[4685]: I1013 08:44:53.618420 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:53 crc kubenswrapper[4685]: I1013 08:44:53.619253 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:53 crc kubenswrapper[4685]: I1013 08:44:53.619314 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:53 crc kubenswrapper[4685]: I1013 08:44:53.619327 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.073095 4685 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.458284 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.458474 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.459509 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.459558 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.459569 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.475894 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 13 08:44:54 crc kubenswrapper[4685]: E1013 08:44:54.509117 4685 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.511963 4685 trace.go:236] Trace[706770948]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (13-Oct-2025 08:44:42.308) (total time: 12203ms): Oct 13 08:44:54 crc kubenswrapper[4685]: Trace[706770948]: ---"Objects listed" error: 12203ms (08:44:54.511) Oct 13 08:44:54 crc kubenswrapper[4685]: Trace[706770948]: [12.203873176s] [12.203873176s] END Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.511991 4685 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.512052 4685 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 13 08:44:54 crc kubenswrapper[4685]: E1013 08:44:54.512753 4685 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.513904 4685 trace.go:236] Trace[712044387]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (13-Oct-2025 08:44:42.566) (total time: 11947ms): Oct 13 08:44:54 crc kubenswrapper[4685]: Trace[712044387]: ---"Objects listed" error: 11947ms (08:44:54.513) Oct 13 08:44:54 crc kubenswrapper[4685]: Trace[712044387]: [11.947769954s] [11.947769954s] END Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.513936 4685 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.549868 4685 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.594004 4685 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:39726->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.594059 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:39726->192.168.126.11:17697: read: connection reset by peer" Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.594455 4685 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": EOF" start-of-body= Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.594593 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": EOF" Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.594984 4685 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.595991 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.622517 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.624468 4685 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f" exitCode=255 Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.624534 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f"} Oct 13 08:44:54 crc kubenswrapper[4685]: I1013 08:44:54.687874 4685 scope.go:117] "RemoveContainer" containerID="16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.425991 4685 apiserver.go:52] "Watching apiserver" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.430781 4685 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.431149 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.431527 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.431527 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.431578 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.431665 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.431665 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 13 08:44:55 crc kubenswrapper[4685]: E1013 08:44:55.431723 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:44:55 crc kubenswrapper[4685]: E1013 08:44:55.431773 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.432049 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:44:55 crc kubenswrapper[4685]: E1013 08:44:55.432084 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.434624 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.434760 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.434895 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.435047 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.435175 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.435337 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.435419 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.435444 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.435943 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.442549 4685 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456079 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456107 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456125 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456141 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456157 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456172 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456186 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456200 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456217 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456234 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456258 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456273 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456287 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456305 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456319 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456335 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456350 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456369 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456384 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456398 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456412 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456426 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456441 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456458 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456472 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456492 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456510 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456530 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456549 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456595 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456615 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456636 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456657 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456679 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456694 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456711 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456730 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456752 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456770 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456775 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456793 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456815 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456834 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456853 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456867 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456883 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.456969 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.457187 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.457283 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.457352 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.457413 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.457429 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.457502 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.457513 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.457725 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.457816 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.457855 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.457960 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.458003 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.458183 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.458326 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.458476 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.458487 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.458522 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.458685 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.458693 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.458852 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.458985 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.459022 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.459108 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.459221 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.459303 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.459389 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.459591 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.460690 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.461220 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.461275 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.461317 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.461527 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.461539 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462133 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462143 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462219 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462251 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462287 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462311 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462333 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462354 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462380 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462403 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462428 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462451 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462473 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462490 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462503 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462552 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462577 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462559 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462601 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462646 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462669 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462690 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462713 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462734 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462841 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462867 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462867 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462889 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462928 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462952 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462972 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.462993 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463111 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463119 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463135 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463158 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463178 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463199 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463224 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463247 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463268 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463291 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463313 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463336 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463359 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463384 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463405 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463427 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463449 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463470 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463492 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463515 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463565 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463566 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463578 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463586 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463621 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463649 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463678 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463700 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463720 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463740 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463762 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463786 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463807 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463831 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463852 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463864 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463872 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463893 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463932 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463955 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.463977 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.464000 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.464013 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.464024 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.464048 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.464070 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.464074 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.464088 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.464251 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.464278 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.464468 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.464559 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.464577 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.464644 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.464663 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.464786 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.464897 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.464987 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.464900 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.465133 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.465221 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.465261 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.464092 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.465501 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.465520 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.465537 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.465554 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.465570 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.465591 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.465612 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.467617 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.467902 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.467798 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.469009 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.469241 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.469454 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.469470 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.469716 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.469779 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.470113 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.470152 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.470318 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.470363 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.470649 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.470671 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.470866 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.470886 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.473232 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.473418 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.473690 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.473814 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.473881 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.474132 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.474547 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.474604 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.475537 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.475595 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.475820 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.475839 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: E1013 08:44:55.475995 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:44:55.975973326 +0000 UTC m=+21.123849087 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.476297 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.476483 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.476516 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.476709 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477040 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.476710 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477180 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477198 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477215 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477251 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477280 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477342 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477374 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477407 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477434 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477463 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477490 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477514 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477541 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477571 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477574 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477636 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477685 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477720 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477745 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477771 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477809 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477832 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477840 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477863 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477893 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477943 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477967 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477989 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.477994 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.478037 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.478165 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.478361 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.479451 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.480045 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.480048 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.480099 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.480292 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.480318 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.480333 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.480070 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.480381 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.480670 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.480676 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.480700 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.480795 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.480905 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.481003 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.481029 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.481125 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.481147 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.481440 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.481480 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.481535 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.481631 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.481673 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.481704 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.481728 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.481750 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.481773 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.481797 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.481817 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.481838 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.481856 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.481858 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.481894 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.481896 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.481930 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.481978 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.482228 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.482616 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.482727 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.482178 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483068 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483099 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483122 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483143 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483164 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483186 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483210 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483233 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483255 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483280 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483301 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483325 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483346 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483369 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483391 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483413 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483400 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483443 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483464 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483485 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483508 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483528 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483552 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483574 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483596 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483619 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483641 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483663 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483685 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483707 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483727 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483774 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483806 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483832 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483854 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483877 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483901 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.484038 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.484066 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483074 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483137 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483185 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483372 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.484818 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483596 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.483819 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.484111 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.484760 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.484856 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.485066 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.485161 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.485193 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.485220 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.485226 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.485496 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.485246 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.485876 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.485984 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.486690 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.487026 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.488632 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.488737 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.488766 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.489111 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.489192 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.489408 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.489433 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.489617 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.489797 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.490028 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.490068 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.490240 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.490564 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.490769 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.490887 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.491062 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.491624 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.491681 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.491803 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.491928 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.491930 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.492219 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.492323 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.492549 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.494025 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.494054 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.494064 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.494140 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.494179 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.494303 4685 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.494327 4685 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.494339 4685 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.494351 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.496793 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.497097 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.497133 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: E1013 08:44:55.497296 4685 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.497598 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.497804 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.498148 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.498375 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.498722 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.499083 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.499275 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.499421 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.499686 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.499741 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.499808 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.499942 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: E1013 08:44:55.499947 4685 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500067 4685 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500070 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: E1013 08:44:55.500185 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 08:44:56.000160694 +0000 UTC m=+21.148036455 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 08:44:55 crc kubenswrapper[4685]: E1013 08:44:55.500239 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 08:44:56.000225647 +0000 UTC m=+21.148101408 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500513 4685 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500536 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500550 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500564 4685 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500562 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500575 4685 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500601 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500614 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500625 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500637 4685 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500648 4685 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500659 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500693 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500715 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500727 4685 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500738 4685 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500750 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500761 4685 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500773 4685 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500785 4685 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500812 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500826 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500842 4685 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500853 4685 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500865 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500877 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500889 4685 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500902 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500932 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500945 4685 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500958 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500972 4685 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500986 4685 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.500997 4685 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501013 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501024 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501036 4685 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501050 4685 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501060 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501072 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501083 4685 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501086 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501098 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501241 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501248 4685 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501289 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501303 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501317 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501330 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501342 4685 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501354 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501364 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501374 4685 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501386 4685 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501396 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501407 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501410 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501417 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501471 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501484 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501497 4685 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501511 4685 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501523 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501539 4685 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501553 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501566 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501586 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501599 4685 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501611 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501623 4685 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501638 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501652 4685 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501664 4685 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501676 4685 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501687 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501699 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501711 4685 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501724 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501738 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501750 4685 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501764 4685 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501776 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501790 4685 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501801 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501813 4685 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501825 4685 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501839 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501851 4685 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501863 4685 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501875 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501887 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501900 4685 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.501938 4685 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.503573 4685 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504000 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504033 4685 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504047 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504064 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504080 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504096 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504112 4685 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504244 4685 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504269 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504285 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504301 4685 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504318 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504335 4685 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504352 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504370 4685 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504386 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504403 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504417 4685 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504432 4685 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504448 4685 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504465 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504484 4685 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504502 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504524 4685 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504541 4685 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504553 4685 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504566 4685 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504581 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504597 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504615 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504632 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504649 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504669 4685 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504685 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504701 4685 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504725 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504746 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504765 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504781 4685 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504796 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504811 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504829 4685 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504846 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504867 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504886 4685 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504902 4685 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504944 4685 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504964 4685 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.504981 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505006 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505023 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505041 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505068 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505086 4685 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505103 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505120 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505135 4685 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505150 4685 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505165 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505182 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505212 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505228 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505243 4685 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505258 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505274 4685 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505293 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505307 4685 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505321 4685 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505336 4685 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505351 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505373 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505394 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.505410 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.508036 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.509564 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.510731 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.512018 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.516363 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.516465 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.522995 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.527894 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.532407 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.532436 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.532819 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.533344 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.534192 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.534768 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.535497 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.536604 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.537302 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.537760 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.543180 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.544840 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.545618 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.546976 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: E1013 08:44:55.547618 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 08:44:55 crc kubenswrapper[4685]: E1013 08:44:55.547657 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 08:44:55 crc kubenswrapper[4685]: E1013 08:44:55.547672 4685 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:44:55 crc kubenswrapper[4685]: E1013 08:44:55.547731 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-13 08:44:56.047713169 +0000 UTC m=+21.195588930 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.547797 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.548893 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: E1013 08:44:55.549241 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 08:44:55 crc kubenswrapper[4685]: E1013 08:44:55.549280 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 08:44:55 crc kubenswrapper[4685]: E1013 08:44:55.549299 4685 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:44:55 crc kubenswrapper[4685]: E1013 08:44:55.549366 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-13 08:44:56.049343938 +0000 UTC m=+21.197219759 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.550090 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.550716 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.551396 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.552745 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.553416 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.554866 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.555548 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.556823 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.557008 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.557241 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.557343 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.558383 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.558961 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.560139 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.560635 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.562011 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.562576 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.563584 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.563701 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.564963 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.565465 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.566814 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.568391 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.569273 4685 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.569388 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.571282 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.572401 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.572786 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.574496 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.575473 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.576738 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.577933 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.578407 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.580125 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.583369 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.584784 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.586145 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.595344 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609322 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609419 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609454 4685 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609467 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609480 4685 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609492 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609503 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609513 4685 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609523 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609533 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609543 4685 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609553 4685 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609563 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609574 4685 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609584 4685 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609597 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609607 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609619 4685 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609631 4685 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609644 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609657 4685 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609669 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609681 4685 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609692 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609702 4685 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609712 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609722 4685 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609732 4685 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.609897 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.610010 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.615327 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.615797 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.616757 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.617569 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.618824 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.618964 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.619334 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.620173 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.620638 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.621361 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.625417 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.625895 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.629622 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.636106 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50"} Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.636804 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.650242 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.674350 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.688657 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.702430 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.716330 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.730707 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.745215 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.747607 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.751992 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.766521 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.776482 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.790977 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.817072 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.829307 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.851740 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.863823 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.878290 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.890154 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.899828 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.916662 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.927730 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:55 crc kubenswrapper[4685]: I1013 08:44:55.943665 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.013123 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.013216 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.013265 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:44:56 crc kubenswrapper[4685]: E1013 08:44:56.013368 4685 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 08:44:56 crc kubenswrapper[4685]: E1013 08:44:56.013377 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:44:57.013328361 +0000 UTC m=+22.161204132 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:44:56 crc kubenswrapper[4685]: E1013 08:44:56.013439 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 08:44:57.013420534 +0000 UTC m=+22.161296305 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 08:44:56 crc kubenswrapper[4685]: E1013 08:44:56.013441 4685 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 08:44:56 crc kubenswrapper[4685]: E1013 08:44:56.013544 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 08:44:57.013516828 +0000 UTC m=+22.161392779 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.114040 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.114102 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:44:56 crc kubenswrapper[4685]: E1013 08:44:56.114214 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 08:44:56 crc kubenswrapper[4685]: E1013 08:44:56.114229 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 08:44:56 crc kubenswrapper[4685]: E1013 08:44:56.114239 4685 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:44:56 crc kubenswrapper[4685]: E1013 08:44:56.114307 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 08:44:56 crc kubenswrapper[4685]: E1013 08:44:56.114351 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 08:44:56 crc kubenswrapper[4685]: E1013 08:44:56.114366 4685 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:44:56 crc kubenswrapper[4685]: E1013 08:44:56.114328 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-13 08:44:57.114315614 +0000 UTC m=+22.262191365 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:44:56 crc kubenswrapper[4685]: E1013 08:44:56.114458 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-13 08:44:57.114434289 +0000 UTC m=+22.262310150 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.502524 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:44:56 crc kubenswrapper[4685]: E1013 08:44:56.502638 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.637605 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"ac121bbc0cb4b3534c439740501e65b7a55f03d6fc16d03b584c7e0671a42dab"} Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.638871 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed"} Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.638983 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"4e9a96626a7918750adc18108b61bb29bcde6683ab8c924482796ac728cfe79a"} Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.640215 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc"} Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.640357 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9"} Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.641826 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3bed7ba0f6e384879c06c4a64bbfa31c3d7a4f1c1beab54a8a3e94b860e50d37"} Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.661390 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-cs95z"] Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.661947 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-cs95z" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.664195 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.665385 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.665773 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.689779 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.707153 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.723787 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.754182 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.797275 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.812246 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.821310 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b4636423-2e13-4c88-90fc-48972c3e2ea2-hosts-file\") pod \"node-resolver-cs95z\" (UID: \"b4636423-2e13-4c88-90fc-48972c3e2ea2\") " pod="openshift-dns/node-resolver-cs95z" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.821360 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rclvm\" (UniqueName: \"kubernetes.io/projected/b4636423-2e13-4c88-90fc-48972c3e2ea2-kube-api-access-rclvm\") pod \"node-resolver-cs95z\" (UID: \"b4636423-2e13-4c88-90fc-48972c3e2ea2\") " pod="openshift-dns/node-resolver-cs95z" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.831364 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.855788 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.898119 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.918757 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.922061 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b4636423-2e13-4c88-90fc-48972c3e2ea2-hosts-file\") pod \"node-resolver-cs95z\" (UID: \"b4636423-2e13-4c88-90fc-48972c3e2ea2\") " pod="openshift-dns/node-resolver-cs95z" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.922243 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rclvm\" (UniqueName: \"kubernetes.io/projected/b4636423-2e13-4c88-90fc-48972c3e2ea2-kube-api-access-rclvm\") pod \"node-resolver-cs95z\" (UID: \"b4636423-2e13-4c88-90fc-48972c3e2ea2\") " pod="openshift-dns/node-resolver-cs95z" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.922253 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b4636423-2e13-4c88-90fc-48972c3e2ea2-hosts-file\") pod \"node-resolver-cs95z\" (UID: \"b4636423-2e13-4c88-90fc-48972c3e2ea2\") " pod="openshift-dns/node-resolver-cs95z" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.935200 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.941648 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rclvm\" (UniqueName: \"kubernetes.io/projected/b4636423-2e13-4c88-90fc-48972c3e2ea2-kube-api-access-rclvm\") pod \"node-resolver-cs95z\" (UID: \"b4636423-2e13-4c88-90fc-48972c3e2ea2\") " pod="openshift-dns/node-resolver-cs95z" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.960177 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.973205 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-cs95z" Oct 13 08:44:56 crc kubenswrapper[4685]: I1013 08:44:56.977845 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:56 crc kubenswrapper[4685]: W1013 08:44:56.983447 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4636423_2e13_4c88_90fc_48972c3e2ea2.slice/crio-1da1f6b8633f702e5ed9f5b5c98360d1cefb263ede1056c08b3ea9ca98d4c531 WatchSource:0}: Error finding container 1da1f6b8633f702e5ed9f5b5c98360d1cefb263ede1056c08b3ea9ca98d4c531: Status 404 returned error can't find the container with id 1da1f6b8633f702e5ed9f5b5c98360d1cefb263ede1056c08b3ea9ca98d4c531 Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.006759 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.023610 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.023685 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.023713 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:44:57 crc kubenswrapper[4685]: E1013 08:44:57.023818 4685 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 08:44:57 crc kubenswrapper[4685]: E1013 08:44:57.023867 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 08:44:59.02385451 +0000 UTC m=+24.171730271 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 08:44:57 crc kubenswrapper[4685]: E1013 08:44:57.024130 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:44:59.024120831 +0000 UTC m=+24.171996592 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:44:57 crc kubenswrapper[4685]: E1013 08:44:57.024248 4685 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 08:44:57 crc kubenswrapper[4685]: E1013 08:44:57.024409 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 08:44:59.02438255 +0000 UTC m=+24.172258311 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.054189 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.087088 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.090125 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-hg77k"] Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.090422 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.090864 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-xrvp5"] Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.091388 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.093676 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-zmbv2"] Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.094033 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.094199 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.094577 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.094805 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.095616 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.095786 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.096203 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.096226 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.096297 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.099693 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.099849 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.099972 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.099975 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.123560 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.124021 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.124059 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:44:57 crc kubenswrapper[4685]: E1013 08:44:57.124182 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 08:44:57 crc kubenswrapper[4685]: E1013 08:44:57.124213 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 08:44:57 crc kubenswrapper[4685]: E1013 08:44:57.124183 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 08:44:57 crc kubenswrapper[4685]: E1013 08:44:57.124225 4685 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:44:57 crc kubenswrapper[4685]: E1013 08:44:57.124237 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 08:44:57 crc kubenswrapper[4685]: E1013 08:44:57.124248 4685 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:44:57 crc kubenswrapper[4685]: E1013 08:44:57.124274 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-13 08:44:59.124262073 +0000 UTC m=+24.272137834 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:44:57 crc kubenswrapper[4685]: E1013 08:44:57.124294 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-13 08:44:59.124281674 +0000 UTC m=+24.272157435 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.146866 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.171458 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.192814 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.223328 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.224785 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-host-var-lib-kubelet\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.224863 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-host-run-k8s-cni-cncf-io\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.224902 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-system-cni-dir\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.224949 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-host-run-netns\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.224978 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-etc-kubernetes\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.225004 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdr66\" (UniqueName: \"kubernetes.io/projected/ad75c066-910c-49dc-8e8f-94fe04de919d-kube-api-access-qdr66\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.225049 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d48dh\" (UniqueName: \"kubernetes.io/projected/313a0145-dae1-4ae0-a841-f5547a965794-kube-api-access-d48dh\") pod \"multus-additional-cni-plugins-zmbv2\" (UID: \"313a0145-dae1-4ae0-a841-f5547a965794\") " pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.225174 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-multus-cni-dir\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.225271 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-host-var-lib-cni-bin\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.225301 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/505637be-a3fb-4b68-bd17-9f0ed875fb3c-rootfs\") pod \"machine-config-daemon-xrvp5\" (UID: \"505637be-a3fb-4b68-bd17-9f0ed875fb3c\") " pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.225352 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/505637be-a3fb-4b68-bd17-9f0ed875fb3c-proxy-tls\") pod \"machine-config-daemon-xrvp5\" (UID: \"505637be-a3fb-4b68-bd17-9f0ed875fb3c\") " pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.225373 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/505637be-a3fb-4b68-bd17-9f0ed875fb3c-mcd-auth-proxy-config\") pod \"machine-config-daemon-xrvp5\" (UID: \"505637be-a3fb-4b68-bd17-9f0ed875fb3c\") " pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.225397 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/313a0145-dae1-4ae0-a841-f5547a965794-os-release\") pod \"multus-additional-cni-plugins-zmbv2\" (UID: \"313a0145-dae1-4ae0-a841-f5547a965794\") " pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.225430 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/313a0145-dae1-4ae0-a841-f5547a965794-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zmbv2\" (UID: \"313a0145-dae1-4ae0-a841-f5547a965794\") " pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.225446 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-multus-socket-dir-parent\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.225468 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/313a0145-dae1-4ae0-a841-f5547a965794-system-cni-dir\") pod \"multus-additional-cni-plugins-zmbv2\" (UID: \"313a0145-dae1-4ae0-a841-f5547a965794\") " pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.225485 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/313a0145-dae1-4ae0-a841-f5547a965794-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zmbv2\" (UID: \"313a0145-dae1-4ae0-a841-f5547a965794\") " pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.225526 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-hostroot\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.225541 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ad75c066-910c-49dc-8e8f-94fe04de919d-multus-daemon-config\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.225556 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72fpn\" (UniqueName: \"kubernetes.io/projected/505637be-a3fb-4b68-bd17-9f0ed875fb3c-kube-api-access-72fpn\") pod \"machine-config-daemon-xrvp5\" (UID: \"505637be-a3fb-4b68-bd17-9f0ed875fb3c\") " pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.225591 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-os-release\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.225631 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/313a0145-dae1-4ae0-a841-f5547a965794-cnibin\") pod \"multus-additional-cni-plugins-zmbv2\" (UID: \"313a0145-dae1-4ae0-a841-f5547a965794\") " pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.225719 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-host-var-lib-cni-multus\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.225777 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-host-run-multus-certs\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.225858 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ad75c066-910c-49dc-8e8f-94fe04de919d-cni-binary-copy\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.226134 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-multus-conf-dir\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.226217 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/313a0145-dae1-4ae0-a841-f5547a965794-cni-binary-copy\") pod \"multus-additional-cni-plugins-zmbv2\" (UID: \"313a0145-dae1-4ae0-a841-f5547a965794\") " pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.226340 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-cnibin\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.245763 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.265282 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.278667 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.291481 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.302866 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.317422 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.327513 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d48dh\" (UniqueName: \"kubernetes.io/projected/313a0145-dae1-4ae0-a841-f5547a965794-kube-api-access-d48dh\") pod \"multus-additional-cni-plugins-zmbv2\" (UID: \"313a0145-dae1-4ae0-a841-f5547a965794\") " pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.327647 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-multus-cni-dir\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.327677 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-host-var-lib-cni-bin\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.327707 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/505637be-a3fb-4b68-bd17-9f0ed875fb3c-rootfs\") pod \"machine-config-daemon-xrvp5\" (UID: \"505637be-a3fb-4b68-bd17-9f0ed875fb3c\") " pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.327735 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/505637be-a3fb-4b68-bd17-9f0ed875fb3c-proxy-tls\") pod \"machine-config-daemon-xrvp5\" (UID: \"505637be-a3fb-4b68-bd17-9f0ed875fb3c\") " pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.327762 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/505637be-a3fb-4b68-bd17-9f0ed875fb3c-mcd-auth-proxy-config\") pod \"machine-config-daemon-xrvp5\" (UID: \"505637be-a3fb-4b68-bd17-9f0ed875fb3c\") " pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.327795 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/313a0145-dae1-4ae0-a841-f5547a965794-os-release\") pod \"multus-additional-cni-plugins-zmbv2\" (UID: \"313a0145-dae1-4ae0-a841-f5547a965794\") " pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.327824 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/313a0145-dae1-4ae0-a841-f5547a965794-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zmbv2\" (UID: \"313a0145-dae1-4ae0-a841-f5547a965794\") " pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.327852 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-multus-socket-dir-parent\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.327860 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/505637be-a3fb-4b68-bd17-9f0ed875fb3c-rootfs\") pod \"machine-config-daemon-xrvp5\" (UID: \"505637be-a3fb-4b68-bd17-9f0ed875fb3c\") " pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.327879 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/313a0145-dae1-4ae0-a841-f5547a965794-system-cni-dir\") pod \"multus-additional-cni-plugins-zmbv2\" (UID: \"313a0145-dae1-4ae0-a841-f5547a965794\") " pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.327945 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/313a0145-dae1-4ae0-a841-f5547a965794-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zmbv2\" (UID: \"313a0145-dae1-4ae0-a841-f5547a965794\") " pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.327976 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-hostroot\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.327976 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/313a0145-dae1-4ae0-a841-f5547a965794-system-cni-dir\") pod \"multus-additional-cni-plugins-zmbv2\" (UID: \"313a0145-dae1-4ae0-a841-f5547a965794\") " pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.327997 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-multus-cni-dir\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.328008 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ad75c066-910c-49dc-8e8f-94fe04de919d-multus-daemon-config\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.328111 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72fpn\" (UniqueName: \"kubernetes.io/projected/505637be-a3fb-4b68-bd17-9f0ed875fb3c-kube-api-access-72fpn\") pod \"machine-config-daemon-xrvp5\" (UID: \"505637be-a3fb-4b68-bd17-9f0ed875fb3c\") " pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.328137 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/313a0145-dae1-4ae0-a841-f5547a965794-cnibin\") pod \"multus-additional-cni-plugins-zmbv2\" (UID: \"313a0145-dae1-4ae0-a841-f5547a965794\") " pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.328156 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-os-release\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.328191 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-host-var-lib-cni-multus\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.328215 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-host-run-multus-certs\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.328234 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ad75c066-910c-49dc-8e8f-94fe04de919d-cni-binary-copy\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.328250 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-multus-conf-dir\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.328267 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/313a0145-dae1-4ae0-a841-f5547a965794-cni-binary-copy\") pod \"multus-additional-cni-plugins-zmbv2\" (UID: \"313a0145-dae1-4ae0-a841-f5547a965794\") " pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.328283 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-cnibin\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.328318 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-host-var-lib-kubelet\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.328339 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-host-run-k8s-cni-cncf-io\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.328359 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-etc-kubernetes\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.328377 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdr66\" (UniqueName: \"kubernetes.io/projected/ad75c066-910c-49dc-8e8f-94fe04de919d-kube-api-access-qdr66\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.328396 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-system-cni-dir\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.328414 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-host-run-netns\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.328479 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-host-run-netns\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.328690 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-hostroot\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.328835 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ad75c066-910c-49dc-8e8f-94fe04de919d-multus-daemon-config\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.328832 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-multus-socket-dir-parent\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.328887 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-host-var-lib-cni-bin\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.328965 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-host-run-k8s-cni-cncf-io\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.329022 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-cnibin\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.329063 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-host-var-lib-kubelet\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.329105 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-host-var-lib-cni-multus\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.329147 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/313a0145-dae1-4ae0-a841-f5547a965794-cnibin\") pod \"multus-additional-cni-plugins-zmbv2\" (UID: \"313a0145-dae1-4ae0-a841-f5547a965794\") " pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.329212 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-os-release\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.329459 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/313a0145-dae1-4ae0-a841-f5547a965794-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zmbv2\" (UID: \"313a0145-dae1-4ae0-a841-f5547a965794\") " pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.329525 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/313a0145-dae1-4ae0-a841-f5547a965794-cni-binary-copy\") pod \"multus-additional-cni-plugins-zmbv2\" (UID: \"313a0145-dae1-4ae0-a841-f5547a965794\") " pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.329573 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-host-run-multus-certs\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.329533 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/505637be-a3fb-4b68-bd17-9f0ed875fb3c-mcd-auth-proxy-config\") pod \"machine-config-daemon-xrvp5\" (UID: \"505637be-a3fb-4b68-bd17-9f0ed875fb3c\") " pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.329628 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-etc-kubernetes\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.329652 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-multus-conf-dir\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.329652 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ad75c066-910c-49dc-8e8f-94fe04de919d-system-cni-dir\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.329565 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/313a0145-dae1-4ae0-a841-f5547a965794-os-release\") pod \"multus-additional-cni-plugins-zmbv2\" (UID: \"313a0145-dae1-4ae0-a841-f5547a965794\") " pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.329819 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/313a0145-dae1-4ae0-a841-f5547a965794-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zmbv2\" (UID: \"313a0145-dae1-4ae0-a841-f5547a965794\") " pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.330030 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ad75c066-910c-49dc-8e8f-94fe04de919d-cni-binary-copy\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.337553 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/505637be-a3fb-4b68-bd17-9f0ed875fb3c-proxy-tls\") pod \"machine-config-daemon-xrvp5\" (UID: \"505637be-a3fb-4b68-bd17-9f0ed875fb3c\") " pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.340244 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.345679 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d48dh\" (UniqueName: \"kubernetes.io/projected/313a0145-dae1-4ae0-a841-f5547a965794-kube-api-access-d48dh\") pod \"multus-additional-cni-plugins-zmbv2\" (UID: \"313a0145-dae1-4ae0-a841-f5547a965794\") " pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.349392 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdr66\" (UniqueName: \"kubernetes.io/projected/ad75c066-910c-49dc-8e8f-94fe04de919d-kube-api-access-qdr66\") pod \"multus-hg77k\" (UID: \"ad75c066-910c-49dc-8e8f-94fe04de919d\") " pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.355706 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.359495 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72fpn\" (UniqueName: \"kubernetes.io/projected/505637be-a3fb-4b68-bd17-9f0ed875fb3c-kube-api-access-72fpn\") pod \"machine-config-daemon-xrvp5\" (UID: \"505637be-a3fb-4b68-bd17-9f0ed875fb3c\") " pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.402550 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-hg77k" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.411026 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 08:44:57 crc kubenswrapper[4685]: W1013 08:44:57.417100 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad75c066_910c_49dc_8e8f_94fe04de919d.slice/crio-f8bc7e913624090e33b0b44f90276728f8d26d3906d7e34087d702455d0f9dc0 WatchSource:0}: Error finding container f8bc7e913624090e33b0b44f90276728f8d26d3906d7e34087d702455d0f9dc0: Status 404 returned error can't find the container with id f8bc7e913624090e33b0b44f90276728f8d26d3906d7e34087d702455d0f9dc0 Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.421837 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" Oct 13 08:44:57 crc kubenswrapper[4685]: W1013 08:44:57.434634 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod505637be_a3fb_4b68_bd17_9f0ed875fb3c.slice/crio-32b5f10d8293d8b58ad2ea8c901869c32a9f1b3b923d7a64195a24769a08ffb1 WatchSource:0}: Error finding container 32b5f10d8293d8b58ad2ea8c901869c32a9f1b3b923d7a64195a24769a08ffb1: Status 404 returned error can't find the container with id 32b5f10d8293d8b58ad2ea8c901869c32a9f1b3b923d7a64195a24769a08ffb1 Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.502541 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:44:57 crc kubenswrapper[4685]: E1013 08:44:57.502683 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.503252 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:44:57 crc kubenswrapper[4685]: E1013 08:44:57.503319 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.521486 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5kgxk"] Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.522422 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.526641 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.528398 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.528465 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.528884 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.528950 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.529058 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.529218 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.546266 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.581645 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.624232 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.632036 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-var-lib-openvswitch\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.632077 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-log-socket\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.632098 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-slash\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.632115 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-node-log\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.632131 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-cni-bin\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.632146 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-run-openvswitch\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.632161 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-run-ovn\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.632181 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.632289 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-run-systemd\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.632332 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-run-ovn-kubernetes\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.632360 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4hsd\" (UniqueName: \"kubernetes.io/projected/ffa4857b-f741-47d8-b692-74a84be30411-kube-api-access-k4hsd\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.632388 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-etc-openvswitch\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.632405 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-cni-netd\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.632424 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ffa4857b-f741-47d8-b692-74a84be30411-env-overrides\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.632456 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ffa4857b-f741-47d8-b692-74a84be30411-ovnkube-script-lib\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.632474 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-kubelet\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.632509 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ffa4857b-f741-47d8-b692-74a84be30411-ovn-node-metrics-cert\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.632541 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-systemd-units\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.632565 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-run-netns\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.632579 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ffa4857b-f741-47d8-b692-74a84be30411-ovnkube-config\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.656818 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.657874 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" event={"ID":"313a0145-dae1-4ae0-a841-f5547a965794","Type":"ContainerStarted","Data":"b7c93d4f4fc086db30671271e488970af7560eeb5a171f66d5b980edcf819d4c"} Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.661653 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerStarted","Data":"d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008"} Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.661684 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerStarted","Data":"32b5f10d8293d8b58ad2ea8c901869c32a9f1b3b923d7a64195a24769a08ffb1"} Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.667377 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hg77k" event={"ID":"ad75c066-910c-49dc-8e8f-94fe04de919d","Type":"ContainerStarted","Data":"a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234"} Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.667454 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hg77k" event={"ID":"ad75c066-910c-49dc-8e8f-94fe04de919d","Type":"ContainerStarted","Data":"f8bc7e913624090e33b0b44f90276728f8d26d3906d7e34087d702455d0f9dc0"} Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.669939 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-cs95z" event={"ID":"b4636423-2e13-4c88-90fc-48972c3e2ea2","Type":"ContainerStarted","Data":"b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60"} Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.670000 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-cs95z" event={"ID":"b4636423-2e13-4c88-90fc-48972c3e2ea2","Type":"ContainerStarted","Data":"1da1f6b8633f702e5ed9f5b5c98360d1cefb263ede1056c08b3ea9ca98d4c531"} Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.691621 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.716043 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.733731 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ffa4857b-f741-47d8-b692-74a84be30411-ovn-node-metrics-cert\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.733774 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-systemd-units\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.733803 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ffa4857b-f741-47d8-b692-74a84be30411-ovnkube-config\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.733821 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-run-netns\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.733856 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-var-lib-openvswitch\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.733880 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-log-socket\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.733897 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-cni-bin\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.733936 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-slash\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.733956 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-node-log\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.733933 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.734031 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.734082 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-log-socket\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.733976 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.734137 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-var-lib-openvswitch\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.734250 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-run-openvswitch\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.734275 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-run-ovn\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.734306 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-run-systemd\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.734359 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-run-ovn\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.734391 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-run-ovn-kubernetes\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.734412 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4hsd\" (UniqueName: \"kubernetes.io/projected/ffa4857b-f741-47d8-b692-74a84be30411-kube-api-access-k4hsd\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.734434 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-etc-openvswitch\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.734454 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-cni-netd\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.734484 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-run-netns\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.734487 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ffa4857b-f741-47d8-b692-74a84be30411-env-overrides\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.734525 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ffa4857b-f741-47d8-b692-74a84be30411-ovnkube-script-lib\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.734543 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-kubelet\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.734600 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-run-systemd\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.734616 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-cni-bin\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.734632 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-slash\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.734647 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-node-log\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.734663 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-systemd-units\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.735162 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ffa4857b-f741-47d8-b692-74a84be30411-env-overrides\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.735168 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-run-ovn-kubernetes\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.735365 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ffa4857b-f741-47d8-b692-74a84be30411-ovnkube-config\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.735485 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-cni-netd\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.735520 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-run-openvswitch\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.735714 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-etc-openvswitch\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.735745 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-kubelet\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.736324 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ffa4857b-f741-47d8-b692-74a84be30411-ovnkube-script-lib\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.741290 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ffa4857b-f741-47d8-b692-74a84be30411-ovn-node-metrics-cert\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.761854 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.764137 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4hsd\" (UniqueName: \"kubernetes.io/projected/ffa4857b-f741-47d8-b692-74a84be30411-kube-api-access-k4hsd\") pod \"ovnkube-node-5kgxk\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.777259 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.791880 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.803832 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.821602 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.838741 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.839135 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: W1013 08:44:57.858186 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podffa4857b_f741_47d8_b692_74a84be30411.slice/crio-c6987d47cba056a20939192576b5ffc101a90da52ee501b1adaa15f600358ddb WatchSource:0}: Error finding container c6987d47cba056a20939192576b5ffc101a90da52ee501b1adaa15f600358ddb: Status 404 returned error can't find the container with id c6987d47cba056a20939192576b5ffc101a90da52ee501b1adaa15f600358ddb Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.871095 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.910768 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.938655 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.957464 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:57 crc kubenswrapper[4685]: I1013 08:44:57.986081 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:57Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.020237 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.065778 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.109396 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.130813 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.146043 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.162223 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.179300 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.194865 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.502110 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:44:58 crc kubenswrapper[4685]: E1013 08:44:58.502264 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.674857 4685 generic.go:334] "Generic (PLEG): container finished" podID="313a0145-dae1-4ae0-a841-f5547a965794" containerID="79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c" exitCode=0 Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.674986 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" event={"ID":"313a0145-dae1-4ae0-a841-f5547a965794","Type":"ContainerDied","Data":"79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c"} Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.679330 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerStarted","Data":"6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9"} Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.683252 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382"} Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.685617 4685 generic.go:334] "Generic (PLEG): container finished" podID="ffa4857b-f741-47d8-b692-74a84be30411" containerID="0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5" exitCode=0 Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.685650 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerDied","Data":"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5"} Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.685670 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerStarted","Data":"c6987d47cba056a20939192576b5ffc101a90da52ee501b1adaa15f600358ddb"} Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.701234 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.720328 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.747575 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.778555 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.794141 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.814433 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.835392 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.853293 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.865242 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.883674 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.897541 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.913617 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.928239 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.948631 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.958989 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.972167 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:58 crc kubenswrapper[4685]: I1013 08:44:58.997113 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:58Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.012466 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.028147 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.048348 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.048511 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.048548 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.048480 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: E1013 08:44:59.048675 4685 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 08:44:59 crc kubenswrapper[4685]: E1013 08:44:59.048737 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 08:45:03.04872014 +0000 UTC m=+28.196595901 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 08:44:59 crc kubenswrapper[4685]: E1013 08:44:59.048762 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:45:03.048750871 +0000 UTC m=+28.196626872 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:44:59 crc kubenswrapper[4685]: E1013 08:44:59.048818 4685 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 08:44:59 crc kubenswrapper[4685]: E1013 08:44:59.048876 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 08:45:03.048855685 +0000 UTC m=+28.196731446 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.065086 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.085029 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.098183 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.117509 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.137880 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.149168 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.149230 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:44:59 crc kubenswrapper[4685]: E1013 08:44:59.149406 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 08:44:59 crc kubenswrapper[4685]: E1013 08:44:59.149425 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 08:44:59 crc kubenswrapper[4685]: E1013 08:44:59.149439 4685 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:44:59 crc kubenswrapper[4685]: E1013 08:44:59.149474 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 08:44:59 crc kubenswrapper[4685]: E1013 08:44:59.149523 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 08:44:59 crc kubenswrapper[4685]: E1013 08:44:59.149541 4685 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:44:59 crc kubenswrapper[4685]: E1013 08:44:59.149498 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-13 08:45:03.149480095 +0000 UTC m=+28.297355856 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:44:59 crc kubenswrapper[4685]: E1013 08:44:59.149629 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-13 08:45:03.14960648 +0000 UTC m=+28.297482451 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.154241 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.502165 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.502233 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:44:59 crc kubenswrapper[4685]: E1013 08:44:59.502717 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:44:59 crc kubenswrapper[4685]: E1013 08:44:59.502854 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.698514 4685 generic.go:334] "Generic (PLEG): container finished" podID="313a0145-dae1-4ae0-a841-f5547a965794" containerID="034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936" exitCode=0 Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.698609 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" event={"ID":"313a0145-dae1-4ae0-a841-f5547a965794","Type":"ContainerDied","Data":"034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936"} Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.704741 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerStarted","Data":"c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b"} Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.704796 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerStarted","Data":"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60"} Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.704809 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerStarted","Data":"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840"} Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.704821 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerStarted","Data":"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023"} Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.704833 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerStarted","Data":"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226"} Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.718401 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.734095 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.754124 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.770379 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.783124 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.795432 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.808414 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.814168 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.816057 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.822603 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.832381 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.847483 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.872172 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.891855 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.913672 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.929405 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.955943 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:44:59 crc kubenswrapper[4685]: I1013 08:44:59.982242 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:44:59Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.003734 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.053248 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.073295 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.095796 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.109594 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.127600 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.143385 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.157740 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.183278 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.198217 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.210976 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.225998 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.269703 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-94gwt"] Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.270602 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-94gwt" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.272550 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.273301 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.273602 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.277751 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.300653 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.315088 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.328976 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.344637 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.358904 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.360351 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e-serviceca\") pod \"node-ca-94gwt\" (UID: \"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\") " pod="openshift-image-registry/node-ca-94gwt" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.360412 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e-host\") pod \"node-ca-94gwt\" (UID: \"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\") " pod="openshift-image-registry/node-ca-94gwt" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.360443 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwgtv\" (UniqueName: \"kubernetes.io/projected/b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e-kube-api-access-lwgtv\") pod \"node-ca-94gwt\" (UID: \"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\") " pod="openshift-image-registry/node-ca-94gwt" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.371612 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.388398 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.402633 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.416210 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.430451 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.444294 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.457546 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.461351 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e-serviceca\") pod \"node-ca-94gwt\" (UID: \"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\") " pod="openshift-image-registry/node-ca-94gwt" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.461391 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e-host\") pod \"node-ca-94gwt\" (UID: \"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\") " pod="openshift-image-registry/node-ca-94gwt" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.461419 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwgtv\" (UniqueName: \"kubernetes.io/projected/b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e-kube-api-access-lwgtv\") pod \"node-ca-94gwt\" (UID: \"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\") " pod="openshift-image-registry/node-ca-94gwt" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.461675 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e-host\") pod \"node-ca-94gwt\" (UID: \"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\") " pod="openshift-image-registry/node-ca-94gwt" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.462763 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e-serviceca\") pod \"node-ca-94gwt\" (UID: \"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\") " pod="openshift-image-registry/node-ca-94gwt" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.480723 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.489803 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwgtv\" (UniqueName: \"kubernetes.io/projected/b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e-kube-api-access-lwgtv\") pod \"node-ca-94gwt\" (UID: \"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\") " pod="openshift-image-registry/node-ca-94gwt" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.502352 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:00 crc kubenswrapper[4685]: E1013 08:45:00.502547 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.503643 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.518460 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.613413 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-94gwt" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.710613 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-94gwt" event={"ID":"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e","Type":"ContainerStarted","Data":"c96b4e531d8f819d9d8ebd2e4401dbfd46f881f44c8e95dc7d18d0f471d026ce"} Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.717777 4685 generic.go:334] "Generic (PLEG): container finished" podID="313a0145-dae1-4ae0-a841-f5547a965794" containerID="724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e" exitCode=0 Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.717866 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" event={"ID":"313a0145-dae1-4ae0-a841-f5547a965794","Type":"ContainerDied","Data":"724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e"} Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.727977 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerStarted","Data":"da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2"} Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.732511 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.751609 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.767524 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.789197 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.806800 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.822984 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.855845 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.872237 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.887226 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.902941 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.913000 4685 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.916710 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.916768 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.916790 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.917014 4685 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.925081 4685 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.925418 4685 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.927644 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.927673 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.927686 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.927703 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.927715 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:00Z","lastTransitionTime":"2025-10-13T08:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.938190 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.958168 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: E1013 08:45:00.958514 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.964463 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.964498 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.964508 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.964522 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.964532 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:00Z","lastTransitionTime":"2025-10-13T08:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:00 crc kubenswrapper[4685]: E1013 08:45:00.977947 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.983413 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.983496 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.983515 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.983545 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.983559 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:00Z","lastTransitionTime":"2025-10-13T08:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:00 crc kubenswrapper[4685]: I1013 08:45:00.985384 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:00 crc kubenswrapper[4685]: E1013 08:45:00.996426 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.000349 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.000384 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.000396 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.000411 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.000421 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:01Z","lastTransitionTime":"2025-10-13T08:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.003146 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:01Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.015852 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:01Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:01 crc kubenswrapper[4685]: E1013 08:45:01.024394 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:01Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.030000 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.030058 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.030094 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.030119 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.030135 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:01Z","lastTransitionTime":"2025-10-13T08:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:01 crc kubenswrapper[4685]: E1013 08:45:01.044815 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:01Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:01 crc kubenswrapper[4685]: E1013 08:45:01.044993 4685 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.047069 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.047114 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.047132 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.047157 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.047174 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:01Z","lastTransitionTime":"2025-10-13T08:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.150235 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.150305 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.150329 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.150362 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.150385 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:01Z","lastTransitionTime":"2025-10-13T08:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.253427 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.253484 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.253496 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.253517 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.253530 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:01Z","lastTransitionTime":"2025-10-13T08:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.356625 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.356693 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.356712 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.356746 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.356770 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:01Z","lastTransitionTime":"2025-10-13T08:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.459508 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.459558 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.459570 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.459589 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.459603 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:01Z","lastTransitionTime":"2025-10-13T08:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.501961 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.502091 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:01 crc kubenswrapper[4685]: E1013 08:45:01.502245 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:01 crc kubenswrapper[4685]: E1013 08:45:01.502518 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.562637 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.562731 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.562756 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.562789 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.562812 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:01Z","lastTransitionTime":"2025-10-13T08:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.667122 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.667193 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.667207 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.667233 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.667249 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:01Z","lastTransitionTime":"2025-10-13T08:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.733569 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-94gwt" event={"ID":"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e","Type":"ContainerStarted","Data":"2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf"} Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.737021 4685 generic.go:334] "Generic (PLEG): container finished" podID="313a0145-dae1-4ae0-a841-f5547a965794" containerID="60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea" exitCode=0 Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.737073 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" event={"ID":"313a0145-dae1-4ae0-a841-f5547a965794","Type":"ContainerDied","Data":"60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea"} Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.771302 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:01Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.773984 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.774044 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.774065 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.774092 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.774112 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:01Z","lastTransitionTime":"2025-10-13T08:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.784701 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:01Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.808776 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:01Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.827897 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:01Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.848593 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:01Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.869273 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:01Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.884890 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.885007 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.885026 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.885052 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.885073 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:01Z","lastTransitionTime":"2025-10-13T08:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.897983 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:01Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.923048 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:01Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.939114 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:01Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.951897 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:01Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.966030 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:01Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.982713 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:01Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.987240 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.987309 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.987321 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.987348 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.987367 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:01Z","lastTransitionTime":"2025-10-13T08:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:01 crc kubenswrapper[4685]: I1013 08:45:01.996476 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:01Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.013390 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.027821 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.041659 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.055595 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.074249 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.089926 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.089983 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.089998 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.090017 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.090029 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:02Z","lastTransitionTime":"2025-10-13T08:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.091769 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.110812 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.132282 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.146255 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.161448 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.174135 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.190595 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.191832 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.191872 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.191883 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.191900 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.191929 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:02Z","lastTransitionTime":"2025-10-13T08:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.202663 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.218686 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.230965 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.246735 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.261763 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.295210 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.295550 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.295636 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.295729 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.295810 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:02Z","lastTransitionTime":"2025-10-13T08:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.399251 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.399296 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.399309 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.399343 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.399364 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:02Z","lastTransitionTime":"2025-10-13T08:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.503542 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:02 crc kubenswrapper[4685]: E1013 08:45:02.503982 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.509490 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.510024 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.510249 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.510451 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.510671 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:02Z","lastTransitionTime":"2025-10-13T08:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.622697 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.622767 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.622787 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.622812 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.622826 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:02Z","lastTransitionTime":"2025-10-13T08:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.724923 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.725228 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.725484 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.725574 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.725643 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:02Z","lastTransitionTime":"2025-10-13T08:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.751287 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerStarted","Data":"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73"} Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.752898 4685 generic.go:334] "Generic (PLEG): container finished" podID="313a0145-dae1-4ae0-a841-f5547a965794" containerID="4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d" exitCode=0 Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.753949 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" event={"ID":"313a0145-dae1-4ae0-a841-f5547a965794","Type":"ContainerDied","Data":"4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d"} Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.771452 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.812738 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.862723 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.862778 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.862788 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.862822 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.862834 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:02Z","lastTransitionTime":"2025-10-13T08:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.870287 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.916539 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.934807 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.956245 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.965476 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.965502 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.965511 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.965527 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.965537 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:02Z","lastTransitionTime":"2025-10-13T08:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.976247 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:02 crc kubenswrapper[4685]: I1013 08:45:02.990776 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.011561 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:03Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.027252 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:03Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.043161 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:03Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.057626 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:03Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.067186 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.067650 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.067660 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.067678 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.067689 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:03Z","lastTransitionTime":"2025-10-13T08:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.070891 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:03Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.084765 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:03Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.095889 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:03Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.100295 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.100421 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:03 crc kubenswrapper[4685]: E1013 08:45:03.100472 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:45:11.100451343 +0000 UTC m=+36.248327104 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.100515 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:03 crc kubenswrapper[4685]: E1013 08:45:03.100517 4685 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 08:45:03 crc kubenswrapper[4685]: E1013 08:45:03.100583 4685 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 08:45:03 crc kubenswrapper[4685]: E1013 08:45:03.100601 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 08:45:11.100594707 +0000 UTC m=+36.248470458 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 08:45:03 crc kubenswrapper[4685]: E1013 08:45:03.100621 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 08:45:11.100611527 +0000 UTC m=+36.248487288 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.170030 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.170186 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.170248 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.170317 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.170397 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:03Z","lastTransitionTime":"2025-10-13T08:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.201699 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.201953 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:03 crc kubenswrapper[4685]: E1013 08:45:03.202065 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 08:45:03 crc kubenswrapper[4685]: E1013 08:45:03.202129 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 08:45:03 crc kubenswrapper[4685]: E1013 08:45:03.202145 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 08:45:03 crc kubenswrapper[4685]: E1013 08:45:03.202161 4685 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:45:03 crc kubenswrapper[4685]: E1013 08:45:03.202188 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 08:45:03 crc kubenswrapper[4685]: E1013 08:45:03.202210 4685 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:45:03 crc kubenswrapper[4685]: E1013 08:45:03.202303 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-13 08:45:11.202261751 +0000 UTC m=+36.350137672 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:45:03 crc kubenswrapper[4685]: E1013 08:45:03.202353 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-13 08:45:11.202335153 +0000 UTC m=+36.350211184 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.273510 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.273791 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.274147 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.274612 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.274894 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:03Z","lastTransitionTime":"2025-10-13T08:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.379037 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.379429 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.379575 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.379676 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.379793 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:03Z","lastTransitionTime":"2025-10-13T08:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.483754 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.483818 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.483829 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.483862 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.483879 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:03Z","lastTransitionTime":"2025-10-13T08:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.502728 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.502782 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:03 crc kubenswrapper[4685]: E1013 08:45:03.503501 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:03 crc kubenswrapper[4685]: E1013 08:45:03.503569 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.587251 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.587298 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.587309 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.587327 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.587340 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:03Z","lastTransitionTime":"2025-10-13T08:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.690521 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.690590 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.690602 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.690621 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.690651 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:03Z","lastTransitionTime":"2025-10-13T08:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.769083 4685 generic.go:334] "Generic (PLEG): container finished" podID="313a0145-dae1-4ae0-a841-f5547a965794" containerID="656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055" exitCode=0 Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.769263 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" event={"ID":"313a0145-dae1-4ae0-a841-f5547a965794","Type":"ContainerDied","Data":"656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055"} Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.796859 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:03Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.800748 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.800808 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.800825 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.800853 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.800869 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:03Z","lastTransitionTime":"2025-10-13T08:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.811381 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:03Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.831364 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:03Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.847312 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:03Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.861839 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:03Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.874454 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:03Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.888824 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:03Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.904059 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:03Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.904731 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.904757 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.904766 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.904783 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.904794 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:03Z","lastTransitionTime":"2025-10-13T08:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.919149 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:03Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.942967 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:03Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.958614 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:03Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.973311 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:03Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:03 crc kubenswrapper[4685]: I1013 08:45:03.985579 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:03Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.006009 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:04Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.009187 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.009250 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.009263 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.009286 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.009302 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:04Z","lastTransitionTime":"2025-10-13T08:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.021550 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:04Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.112831 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.113236 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.113254 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.113277 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.113291 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:04Z","lastTransitionTime":"2025-10-13T08:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.216109 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.216162 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.216178 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.216204 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.216221 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:04Z","lastTransitionTime":"2025-10-13T08:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.319154 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.319207 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.319222 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.319248 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.319263 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:04Z","lastTransitionTime":"2025-10-13T08:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.422722 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.422803 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.422830 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.422865 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.422894 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:04Z","lastTransitionTime":"2025-10-13T08:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.502242 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:04 crc kubenswrapper[4685]: E1013 08:45:04.502383 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.525289 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.525342 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.525352 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.525376 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.525390 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:04Z","lastTransitionTime":"2025-10-13T08:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.659632 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.659681 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.659694 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.659716 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.659729 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:04Z","lastTransitionTime":"2025-10-13T08:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.762635 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.762672 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.762681 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.762696 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.762706 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:04Z","lastTransitionTime":"2025-10-13T08:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.779514 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerStarted","Data":"5e5717d18cb9d9452ab2bd0e1f515e7c88f492fec9a944cb1746d8eea242cefd"} Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.780225 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.791042 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" event={"ID":"313a0145-dae1-4ae0-a841-f5547a965794","Type":"ContainerStarted","Data":"25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb"} Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.808274 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:04Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.813344 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.827593 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:04Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.841200 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:04Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.860788 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:04Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.865942 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.865989 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.866040 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.866069 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.866088 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:04Z","lastTransitionTime":"2025-10-13T08:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.874035 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:04Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.886718 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:04Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.902382 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:04Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.923164 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:04Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.941465 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:04Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.959681 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:04Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.971314 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.971386 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.971402 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.971427 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.971447 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:04Z","lastTransitionTime":"2025-10-13T08:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.979855 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:04Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:04 crc kubenswrapper[4685]: I1013 08:45:04.999020 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:04Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.018088 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.036686 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.061699 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5717d18cb9d9452ab2bd0e1f515e7c88f492fec9a944cb1746d8eea242cefd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.073871 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.073982 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.074082 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.074093 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.074109 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.074119 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:05Z","lastTransitionTime":"2025-10-13T08:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.091844 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.105364 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.123660 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.138234 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.152073 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.168490 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.177133 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.177175 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.177185 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.177200 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.177208 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:05Z","lastTransitionTime":"2025-10-13T08:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.184869 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.196731 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.218363 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.235244 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.250261 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.272149 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5717d18cb9d9452ab2bd0e1f515e7c88f492fec9a944cb1746d8eea242cefd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.279856 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.279904 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.279930 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.279953 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.279971 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:05Z","lastTransitionTime":"2025-10-13T08:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.292338 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.305115 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.383883 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.383962 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.383974 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.383993 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.384005 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:05Z","lastTransitionTime":"2025-10-13T08:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.488552 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.488604 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.488614 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.488634 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.488647 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:05Z","lastTransitionTime":"2025-10-13T08:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.503102 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:05 crc kubenswrapper[4685]: E1013 08:45:05.503289 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.503720 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:05 crc kubenswrapper[4685]: E1013 08:45:05.503828 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.519610 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.530752 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.545314 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.554989 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.569775 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.584584 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.591096 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.591122 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.591132 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.591151 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.591164 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:05Z","lastTransitionTime":"2025-10-13T08:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.603118 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.618397 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.632350 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.646285 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.672935 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.695142 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.695200 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.695215 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.695236 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.695264 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:05Z","lastTransitionTime":"2025-10-13T08:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.698600 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.710863 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.730238 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5717d18cb9d9452ab2bd0e1f515e7c88f492fec9a944cb1746d8eea242cefd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.744251 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.795181 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.795242 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.796950 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.796991 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.797003 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.797022 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.797036 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:05Z","lastTransitionTime":"2025-10-13T08:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.819870 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.841133 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.859530 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.872528 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.886360 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.900539 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.900588 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.900601 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.900623 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.900639 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:05Z","lastTransitionTime":"2025-10-13T08:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.905199 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.920872 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.935397 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.958777 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.975761 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:05 crc kubenswrapper[4685]: I1013 08:45:05.991098 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:05Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.004438 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.004496 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.004513 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.004531 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.004544 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:06Z","lastTransitionTime":"2025-10-13T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.008074 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:06Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.023503 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:06Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.035619 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:06Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.049005 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:06Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.069353 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5717d18cb9d9452ab2bd0e1f515e7c88f492fec9a944cb1746d8eea242cefd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:06Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.107748 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.107798 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.107809 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.107829 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.107841 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:06Z","lastTransitionTime":"2025-10-13T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.210798 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.211160 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.211307 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.211426 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.211503 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:06Z","lastTransitionTime":"2025-10-13T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.314553 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.314993 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.315058 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.315121 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.315626 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:06Z","lastTransitionTime":"2025-10-13T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.419410 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.419732 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.419845 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.419940 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.420021 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:06Z","lastTransitionTime":"2025-10-13T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.501892 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:06 crc kubenswrapper[4685]: E1013 08:45:06.502115 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.523123 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.523176 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.523187 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.523212 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.523227 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:06Z","lastTransitionTime":"2025-10-13T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.625999 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.626047 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.626063 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.626082 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.626093 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:06Z","lastTransitionTime":"2025-10-13T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.728949 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.728999 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.729016 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.729035 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.729047 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:06Z","lastTransitionTime":"2025-10-13T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.832028 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.832065 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.832075 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.832091 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.832103 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:06Z","lastTransitionTime":"2025-10-13T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.934808 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.934847 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.934855 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.934873 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:06 crc kubenswrapper[4685]: I1013 08:45:06.934884 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:06Z","lastTransitionTime":"2025-10-13T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.038391 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.038456 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.038468 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.038537 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.038548 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:07Z","lastTransitionTime":"2025-10-13T08:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.141299 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.141348 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.141357 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.141376 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.141387 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:07Z","lastTransitionTime":"2025-10-13T08:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.244567 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.244636 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.244647 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.244664 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.244679 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:07Z","lastTransitionTime":"2025-10-13T08:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.350873 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.350952 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.350964 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.350992 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.351009 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:07Z","lastTransitionTime":"2025-10-13T08:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.453525 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.453606 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.453720 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.453755 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.453779 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:07Z","lastTransitionTime":"2025-10-13T08:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.502042 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:07 crc kubenswrapper[4685]: E1013 08:45:07.502239 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.502851 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:07 crc kubenswrapper[4685]: E1013 08:45:07.503022 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.557011 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.557239 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.557314 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.557413 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.557486 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:07Z","lastTransitionTime":"2025-10-13T08:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.660694 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.661087 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.661350 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.661637 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.661898 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:07Z","lastTransitionTime":"2025-10-13T08:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.765075 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.765132 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.765145 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.765164 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.765178 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:07Z","lastTransitionTime":"2025-10-13T08:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.805801 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5kgxk_ffa4857b-f741-47d8-b692-74a84be30411/ovnkube-controller/0.log" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.812676 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerDied","Data":"5e5717d18cb9d9452ab2bd0e1f515e7c88f492fec9a944cb1746d8eea242cefd"} Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.813086 4685 generic.go:334] "Generic (PLEG): container finished" podID="ffa4857b-f741-47d8-b692-74a84be30411" containerID="5e5717d18cb9d9452ab2bd0e1f515e7c88f492fec9a944cb1746d8eea242cefd" exitCode=1 Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.814974 4685 scope.go:117] "RemoveContainer" containerID="5e5717d18cb9d9452ab2bd0e1f515e7c88f492fec9a944cb1746d8eea242cefd" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.830424 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:07Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.858011 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:07Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.868998 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.869039 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.869048 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.869066 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.869077 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:07Z","lastTransitionTime":"2025-10-13T08:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.885134 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:07Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.901107 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:07Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.918316 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:07Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.939495 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5717d18cb9d9452ab2bd0e1f515e7c88f492fec9a944cb1746d8eea242cefd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5717d18cb9d9452ab2bd0e1f515e7c88f492fec9a944cb1746d8eea242cefd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:07Z\\\",\\\"message\\\":\\\"or.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 08:45:07.161312 5904 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 08:45:07.161676 5904 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 08:45:07.161953 5904 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 08:45:07.162112 5904 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1013 08:45:07.162396 5904 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1013 08:45:07.162595 5904 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1013 08:45:07.162685 5904 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 08:45:07.163126 5904 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:07Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.957026 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:07Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.973145 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.973179 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.973192 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.973214 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.973227 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:07Z","lastTransitionTime":"2025-10-13T08:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:07 crc kubenswrapper[4685]: I1013 08:45:07.982515 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:07Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.001434 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:07Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.018790 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:08Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.032056 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:08Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.045992 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:08Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.059740 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:08Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.072285 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:08Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.076065 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.076108 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.076122 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.076143 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.076157 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:08Z","lastTransitionTime":"2025-10-13T08:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.084845 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:08Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.180076 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.180118 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.180128 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.180144 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.180157 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:08Z","lastTransitionTime":"2025-10-13T08:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.283191 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.283246 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.283259 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.283282 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.283352 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:08Z","lastTransitionTime":"2025-10-13T08:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.386361 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.386418 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.386453 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.386475 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.386500 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:08Z","lastTransitionTime":"2025-10-13T08:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.489829 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.490241 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.490342 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.490433 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.490503 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:08Z","lastTransitionTime":"2025-10-13T08:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.502230 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:08 crc kubenswrapper[4685]: E1013 08:45:08.502376 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.592971 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.593023 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.593034 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.593051 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.593061 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:08Z","lastTransitionTime":"2025-10-13T08:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.695744 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.696073 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.696141 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.696208 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.696304 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:08Z","lastTransitionTime":"2025-10-13T08:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.798895 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.798974 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.798991 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.799012 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.799027 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:08Z","lastTransitionTime":"2025-10-13T08:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.823386 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5kgxk_ffa4857b-f741-47d8-b692-74a84be30411/ovnkube-controller/0.log" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.827604 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerStarted","Data":"4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe"} Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.828445 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.843430 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:08Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.864576 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:08Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.877223 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:08Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.894512 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:08Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.902139 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.902173 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.902185 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.902217 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.902230 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:08Z","lastTransitionTime":"2025-10-13T08:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.908276 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:08Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.923437 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:08Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.936051 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:08Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.953595 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:08Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.970020 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:08Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:08 crc kubenswrapper[4685]: I1013 08:45:08.995296 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:08Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.004599 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.004631 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.004639 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.004652 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.004660 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:09Z","lastTransitionTime":"2025-10-13T08:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.012381 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:09Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.028795 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:09Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.107292 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.107649 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.107707 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.107788 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.107862 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:09Z","lastTransitionTime":"2025-10-13T08:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.116430 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5717d18cb9d9452ab2bd0e1f515e7c88f492fec9a944cb1746d8eea242cefd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:07Z\\\",\\\"message\\\":\\\"or.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 08:45:07.161312 5904 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 08:45:07.161676 5904 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 08:45:07.161953 5904 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 08:45:07.162112 5904 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1013 08:45:07.162396 5904 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1013 08:45:07.162595 5904 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1013 08:45:07.162685 5904 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 08:45:07.163126 5904 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:09Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.134736 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:09Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.151613 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:09Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.211088 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.211126 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.211139 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.211156 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.211168 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:09Z","lastTransitionTime":"2025-10-13T08:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.313334 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.313369 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.313379 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.313394 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.313405 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:09Z","lastTransitionTime":"2025-10-13T08:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.415424 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.415496 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.415518 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.415547 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.415568 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:09Z","lastTransitionTime":"2025-10-13T08:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.501772 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.501784 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:09 crc kubenswrapper[4685]: E1013 08:45:09.502440 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:09 crc kubenswrapper[4685]: E1013 08:45:09.502636 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.518465 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.518515 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.518532 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.518551 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.518568 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:09Z","lastTransitionTime":"2025-10-13T08:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.621439 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.621480 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.621491 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.621507 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.621519 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:09Z","lastTransitionTime":"2025-10-13T08:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.724342 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.724419 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.724441 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.724471 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.724518 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:09Z","lastTransitionTime":"2025-10-13T08:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.775700 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r"] Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.776421 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.779935 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.781231 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.791036 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:09Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.805014 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295380a1-3f1f-4c5e-907c-b0545aa739e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p8d8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:09Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.817593 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:09Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.817620 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/295380a1-3f1f-4c5e-907c-b0545aa739e2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-p8d8r\" (UID: \"295380a1-3f1f-4c5e-907c-b0545aa739e2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.817768 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/295380a1-3f1f-4c5e-907c-b0545aa739e2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-p8d8r\" (UID: \"295380a1-3f1f-4c5e-907c-b0545aa739e2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.817794 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn5qz\" (UniqueName: \"kubernetes.io/projected/295380a1-3f1f-4c5e-907c-b0545aa739e2-kube-api-access-jn5qz\") pod \"ovnkube-control-plane-749d76644c-p8d8r\" (UID: \"295380a1-3f1f-4c5e-907c-b0545aa739e2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.817841 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/295380a1-3f1f-4c5e-907c-b0545aa739e2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-p8d8r\" (UID: \"295380a1-3f1f-4c5e-907c-b0545aa739e2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.827606 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.827652 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.827664 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.827681 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.827694 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:09Z","lastTransitionTime":"2025-10-13T08:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.832087 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5kgxk_ffa4857b-f741-47d8-b692-74a84be30411/ovnkube-controller/1.log" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.832506 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5kgxk_ffa4857b-f741-47d8-b692-74a84be30411/ovnkube-controller/0.log" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.832970 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:09Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.834987 4685 generic.go:334] "Generic (PLEG): container finished" podID="ffa4857b-f741-47d8-b692-74a84be30411" containerID="4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe" exitCode=1 Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.835024 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerDied","Data":"4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe"} Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.835075 4685 scope.go:117] "RemoveContainer" containerID="5e5717d18cb9d9452ab2bd0e1f515e7c88f492fec9a944cb1746d8eea242cefd" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.835834 4685 scope.go:117] "RemoveContainer" containerID="4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe" Oct 13 08:45:09 crc kubenswrapper[4685]: E1013 08:45:09.836083 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5kgxk_openshift-ovn-kubernetes(ffa4857b-f741-47d8-b692-74a84be30411)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" podUID="ffa4857b-f741-47d8-b692-74a84be30411" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.852012 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:09Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.863321 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:09Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.876304 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:09Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.894691 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5717d18cb9d9452ab2bd0e1f515e7c88f492fec9a944cb1746d8eea242cefd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:07Z\\\",\\\"message\\\":\\\"or.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 08:45:07.161312 5904 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 08:45:07.161676 5904 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 08:45:07.161953 5904 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 08:45:07.162112 5904 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1013 08:45:07.162396 5904 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1013 08:45:07.162595 5904 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1013 08:45:07.162685 5904 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 08:45:07.163126 5904 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:09Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.915797 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:09Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.918427 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/295380a1-3f1f-4c5e-907c-b0545aa739e2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-p8d8r\" (UID: \"295380a1-3f1f-4c5e-907c-b0545aa739e2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.918488 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn5qz\" (UniqueName: \"kubernetes.io/projected/295380a1-3f1f-4c5e-907c-b0545aa739e2-kube-api-access-jn5qz\") pod \"ovnkube-control-plane-749d76644c-p8d8r\" (UID: \"295380a1-3f1f-4c5e-907c-b0545aa739e2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.918572 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/295380a1-3f1f-4c5e-907c-b0545aa739e2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-p8d8r\" (UID: \"295380a1-3f1f-4c5e-907c-b0545aa739e2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.918604 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/295380a1-3f1f-4c5e-907c-b0545aa739e2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-p8d8r\" (UID: \"295380a1-3f1f-4c5e-907c-b0545aa739e2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.919782 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/295380a1-3f1f-4c5e-907c-b0545aa739e2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-p8d8r\" (UID: \"295380a1-3f1f-4c5e-907c-b0545aa739e2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.919903 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/295380a1-3f1f-4c5e-907c-b0545aa739e2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-p8d8r\" (UID: \"295380a1-3f1f-4c5e-907c-b0545aa739e2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.926430 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/295380a1-3f1f-4c5e-907c-b0545aa739e2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-p8d8r\" (UID: \"295380a1-3f1f-4c5e-907c-b0545aa739e2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.929756 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:09Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.930609 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.930639 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.930647 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.930662 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.930673 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:09Z","lastTransitionTime":"2025-10-13T08:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.944668 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:09Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.948601 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn5qz\" (UniqueName: \"kubernetes.io/projected/295380a1-3f1f-4c5e-907c-b0545aa739e2-kube-api-access-jn5qz\") pod \"ovnkube-control-plane-749d76644c-p8d8r\" (UID: \"295380a1-3f1f-4c5e-907c-b0545aa739e2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.957526 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:09Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.968284 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:09Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.983652 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:09Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:09 crc kubenswrapper[4685]: I1013 08:45:09.995375 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:09Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.006109 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.019742 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.032604 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.032634 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.032645 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.032660 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.032672 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:10Z","lastTransitionTime":"2025-10-13T08:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.034781 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.050540 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.068488 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5717d18cb9d9452ab2bd0e1f515e7c88f492fec9a944cb1746d8eea242cefd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:07Z\\\",\\\"message\\\":\\\"or.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 08:45:07.161312 5904 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 08:45:07.161676 5904 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 08:45:07.161953 5904 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 08:45:07.162112 5904 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1013 08:45:07.162396 5904 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1013 08:45:07.162595 5904 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1013 08:45:07.162685 5904 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 08:45:07.163126 5904 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"message\\\":\\\"p:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.138:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97419c58-41c7-41d7-a137-a446f0c7eeb3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1013 08:45:08.975476 6040 factory.go:656] Stopping watch factory\\\\nI1013 08:45:08.975487 6040 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1013 08:45:08.975497 6040 ovnkube.go:599] Stopped ovnkube\\\\nI1013 08:45:08.975484 6040 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-operators]} name:Service_openshift-marketplace/redhat-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.138:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97419c58-41c7-41d7-a137-a446f0c7eeb3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1013 08:45:08.975531 6040 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1013 08:45:08.975648 6040 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.086951 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.099401 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.103539 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.121850 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.134652 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.134677 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.134686 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.134699 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.134708 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:10Z","lastTransitionTime":"2025-10-13T08:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.135876 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.152017 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.166018 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.179949 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.192540 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295380a1-3f1f-4c5e-907c-b0545aa739e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p8d8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.212179 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.226203 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.238700 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.238762 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.238792 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.238801 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.238816 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.238828 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:10Z","lastTransitionTime":"2025-10-13T08:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.251456 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.342485 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.342517 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.342533 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.342550 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.342560 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:10Z","lastTransitionTime":"2025-10-13T08:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.444681 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.444716 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.444725 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.444742 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.444753 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:10Z","lastTransitionTime":"2025-10-13T08:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.501961 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:10 crc kubenswrapper[4685]: E1013 08:45:10.502121 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.547656 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.547694 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.547706 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.547724 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.547736 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:10Z","lastTransitionTime":"2025-10-13T08:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.649698 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.649744 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.649756 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.649775 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.649789 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:10Z","lastTransitionTime":"2025-10-13T08:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.751669 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.751704 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.751715 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.751730 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.751740 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:10Z","lastTransitionTime":"2025-10-13T08:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.839417 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5kgxk_ffa4857b-f741-47d8-b692-74a84be30411/ovnkube-controller/1.log" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.842631 4685 scope.go:117] "RemoveContainer" containerID="4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe" Oct 13 08:45:10 crc kubenswrapper[4685]: E1013 08:45:10.842808 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5kgxk_openshift-ovn-kubernetes(ffa4857b-f741-47d8-b692-74a84be30411)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" podUID="ffa4857b-f741-47d8-b692-74a84be30411" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.844279 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" event={"ID":"295380a1-3f1f-4c5e-907c-b0545aa739e2","Type":"ContainerStarted","Data":"ca7136baf6f1e826c79c8171ce14480489ff2fab3c1918d6d807535624716b19"} Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.844308 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" event={"ID":"295380a1-3f1f-4c5e-907c-b0545aa739e2","Type":"ContainerStarted","Data":"cf1b49a1ee1c45ddd3d90831ec5ea3064b05a7ac65ae67ccdfe3dc25ae0e5c24"} Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.844319 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" event={"ID":"295380a1-3f1f-4c5e-907c-b0545aa739e2","Type":"ContainerStarted","Data":"5eb9b70a83b59f2a9c3d0a5c5d6d10b53455c79fa43bb377ae430c6840ca8cb3"} Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.853679 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.853722 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.853731 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.853749 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.853757 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:10Z","lastTransitionTime":"2025-10-13T08:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.854140 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295380a1-3f1f-4c5e-907c-b0545aa739e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p8d8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.866944 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.876576 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.887660 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-nkxgw"] Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.888117 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:10 crc kubenswrapper[4685]: E1013 08:45:10.888171 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.899512 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.910414 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.919926 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.928784 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.934711 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs\") pod \"network-metrics-daemon-nkxgw\" (UID: \"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\") " pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.934942 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prj95\" (UniqueName: \"kubernetes.io/projected/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-kube-api-access-prj95\") pod \"network-metrics-daemon-nkxgw\" (UID: \"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\") " pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.944824 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.956056 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.956312 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.956342 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.956374 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.956395 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.956409 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:10Z","lastTransitionTime":"2025-10-13T08:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.972449 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:10 crc kubenswrapper[4685]: I1013 08:45:10.991034 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"message\\\":\\\"p:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.138:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97419c58-41c7-41d7-a137-a446f0c7eeb3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1013 08:45:08.975476 6040 factory.go:656] Stopping watch factory\\\\nI1013 08:45:08.975487 6040 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1013 08:45:08.975497 6040 ovnkube.go:599] Stopped ovnkube\\\\nI1013 08:45:08.975484 6040 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-operators]} name:Service_openshift-marketplace/redhat-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.138:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97419c58-41c7-41d7-a137-a446f0c7eeb3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1013 08:45:08.975531 6040 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1013 08:45:08.975648 6040 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5kgxk_openshift-ovn-kubernetes(ffa4857b-f741-47d8-b692-74a84be30411)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.000943 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.013466 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.024798 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.033804 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.036080 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs\") pod \"network-metrics-daemon-nkxgw\" (UID: \"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\") " pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.036136 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prj95\" (UniqueName: \"kubernetes.io/projected/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-kube-api-access-prj95\") pod \"network-metrics-daemon-nkxgw\" (UID: \"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\") " pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.036310 4685 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.036417 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs podName:7cb40c32-d7d4-4a1b-9cda-1de0fab48566 nodeName:}" failed. No retries permitted until 2025-10-13 08:45:11.536396414 +0000 UTC m=+36.684272175 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs") pod "network-metrics-daemon-nkxgw" (UID: "7cb40c32-d7d4-4a1b-9cda-1de0fab48566") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.047949 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.059273 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.059324 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.059341 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.059364 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.059380 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:11Z","lastTransitionTime":"2025-10-13T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.060812 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prj95\" (UniqueName: \"kubernetes.io/projected/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-kube-api-access-prj95\") pod \"network-metrics-daemon-nkxgw\" (UID: \"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\") " pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.065029 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"message\\\":\\\"p:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.138:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97419c58-41c7-41d7-a137-a446f0c7eeb3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1013 08:45:08.975476 6040 factory.go:656] Stopping watch factory\\\\nI1013 08:45:08.975487 6040 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1013 08:45:08.975497 6040 ovnkube.go:599] Stopped ovnkube\\\\nI1013 08:45:08.975484 6040 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-operators]} name:Service_openshift-marketplace/redhat-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.138:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97419c58-41c7-41d7-a137-a446f0c7eeb3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1013 08:45:08.975531 6040 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1013 08:45:08.975648 6040 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5kgxk_openshift-ovn-kubernetes(ffa4857b-f741-47d8-b692-74a84be30411)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.078701 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.090324 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.103901 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.118734 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.128504 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.136856 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.137062 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.137119 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.137262 4685 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.137303 4685 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.137301 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:45:27.137266807 +0000 UTC m=+52.285142568 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.137463 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 08:45:27.137438052 +0000 UTC m=+52.285313843 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.137526 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 08:45:27.137484344 +0000 UTC m=+52.285360125 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.140778 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.153026 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.162342 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.162399 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.162415 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.162437 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.162455 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:11Z","lastTransitionTime":"2025-10-13T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.164169 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.177054 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.191462 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295380a1-3f1f-4c5e-907c-b0545aa739e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf1b49a1ee1c45ddd3d90831ec5ea3064b05a7ac65ae67ccdfe3dc25ae0e5c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7136baf6f1e826c79c8171ce14480489ff2fab3c1918d6d807535624716b19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p8d8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.207231 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nkxgw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nkxgw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.221208 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.234065 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.238207 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.238299 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.238286 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.238394 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.238428 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.238452 4685 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.238461 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.238497 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.238509 4685 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.238509 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-13 08:45:27.238493511 +0000 UTC m=+52.386369272 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.238589 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-13 08:45:27.238569623 +0000 UTC m=+52.386445384 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.260854 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.264575 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.264603 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.264612 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.264628 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.264638 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:11Z","lastTransitionTime":"2025-10-13T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.273138 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.286390 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.296693 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295380a1-3f1f-4c5e-907c-b0545aa739e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf1b49a1ee1c45ddd3d90831ec5ea3064b05a7ac65ae67ccdfe3dc25ae0e5c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7136baf6f1e826c79c8171ce14480489ff2fab3c1918d6d807535624716b19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p8d8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.307668 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nkxgw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nkxgw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.321278 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.332077 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.349323 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.360784 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.367305 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.367342 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.367351 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.367370 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.367379 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:11Z","lastTransitionTime":"2025-10-13T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.371632 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.387076 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.398640 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.410169 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.418516 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.418555 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.418563 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.418579 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.418588 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:11Z","lastTransitionTime":"2025-10-13T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.424161 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.430617 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.434050 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.434104 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.434116 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.434133 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.434144 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:11Z","lastTransitionTime":"2025-10-13T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.441478 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"message\\\":\\\"p:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.138:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97419c58-41c7-41d7-a137-a446f0c7eeb3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1013 08:45:08.975476 6040 factory.go:656] Stopping watch factory\\\\nI1013 08:45:08.975487 6040 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1013 08:45:08.975497 6040 ovnkube.go:599] Stopped ovnkube\\\\nI1013 08:45:08.975484 6040 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-operators]} name:Service_openshift-marketplace/redhat-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.138:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97419c58-41c7-41d7-a137-a446f0c7eeb3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1013 08:45:08.975531 6040 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1013 08:45:08.975648 6040 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5kgxk_openshift-ovn-kubernetes(ffa4857b-f741-47d8-b692-74a84be30411)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.444760 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.449373 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.449656 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.449667 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.449681 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.449691 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:11Z","lastTransitionTime":"2025-10-13T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.453384 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.460436 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.463411 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.463444 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.463456 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.463473 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.463485 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:11Z","lastTransitionTime":"2025-10-13T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.466072 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.475295 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.477633 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.479253 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.479287 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.479298 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.479314 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.479324 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:11Z","lastTransitionTime":"2025-10-13T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.488306 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.492073 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.492193 4685 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.493644 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.493715 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.493729 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.493770 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.493784 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:11Z","lastTransitionTime":"2025-10-13T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.502042 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.502107 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.502168 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.502297 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.502255 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:11Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.541937 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs\") pod \"network-metrics-daemon-nkxgw\" (UID: \"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\") " pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.542107 4685 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 08:45:11 crc kubenswrapper[4685]: E1013 08:45:11.542188 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs podName:7cb40c32-d7d4-4a1b-9cda-1de0fab48566 nodeName:}" failed. No retries permitted until 2025-10-13 08:45:12.542168988 +0000 UTC m=+37.690044739 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs") pod "network-metrics-daemon-nkxgw" (UID: "7cb40c32-d7d4-4a1b-9cda-1de0fab48566") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.596381 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.596508 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.596521 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.596539 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.596549 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:11Z","lastTransitionTime":"2025-10-13T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.700256 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.700336 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.700355 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.700383 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.700401 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:11Z","lastTransitionTime":"2025-10-13T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.803193 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.803321 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.803388 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.803430 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.803498 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:11Z","lastTransitionTime":"2025-10-13T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.907574 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.907633 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.907646 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.907666 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:11 crc kubenswrapper[4685]: I1013 08:45:11.907731 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:11Z","lastTransitionTime":"2025-10-13T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.011144 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.011204 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.011216 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.011235 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.011254 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:12Z","lastTransitionTime":"2025-10-13T08:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.114392 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.114432 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.114449 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.114465 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.114476 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:12Z","lastTransitionTime":"2025-10-13T08:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.222782 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.222831 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.222841 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.222859 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.222871 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:12Z","lastTransitionTime":"2025-10-13T08:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.324873 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.324941 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.324959 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.325010 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.325032 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:12Z","lastTransitionTime":"2025-10-13T08:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.428459 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.428507 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.428519 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.428540 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.428553 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:12Z","lastTransitionTime":"2025-10-13T08:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.502581 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.502651 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:12 crc kubenswrapper[4685]: E1013 08:45:12.502772 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:12 crc kubenswrapper[4685]: E1013 08:45:12.502846 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.531162 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.531200 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.531210 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.531226 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.531238 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:12Z","lastTransitionTime":"2025-10-13T08:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.554339 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs\") pod \"network-metrics-daemon-nkxgw\" (UID: \"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\") " pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:12 crc kubenswrapper[4685]: E1013 08:45:12.554513 4685 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 08:45:12 crc kubenswrapper[4685]: E1013 08:45:12.554650 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs podName:7cb40c32-d7d4-4a1b-9cda-1de0fab48566 nodeName:}" failed. No retries permitted until 2025-10-13 08:45:14.554631478 +0000 UTC m=+39.702507239 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs") pod "network-metrics-daemon-nkxgw" (UID: "7cb40c32-d7d4-4a1b-9cda-1de0fab48566") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.633829 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.634051 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.634075 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.634110 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.634135 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:12Z","lastTransitionTime":"2025-10-13T08:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.737121 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.737202 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.737217 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.737236 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.737299 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:12Z","lastTransitionTime":"2025-10-13T08:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.840329 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.840404 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.840424 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.840493 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.840512 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:12Z","lastTransitionTime":"2025-10-13T08:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.943188 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.943234 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.943244 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.943263 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:12 crc kubenswrapper[4685]: I1013 08:45:12.943274 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:12Z","lastTransitionTime":"2025-10-13T08:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.045990 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.046026 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.046037 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.046053 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.046063 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:13Z","lastTransitionTime":"2025-10-13T08:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.148310 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.148360 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.148372 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.148390 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.148400 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:13Z","lastTransitionTime":"2025-10-13T08:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.251753 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.251793 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.251803 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.251818 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.251828 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:13Z","lastTransitionTime":"2025-10-13T08:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.353736 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.353800 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.353820 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.353844 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.353862 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:13Z","lastTransitionTime":"2025-10-13T08:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.456388 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.456435 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.456452 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.456471 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.456481 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:13Z","lastTransitionTime":"2025-10-13T08:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.502037 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.502058 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:13 crc kubenswrapper[4685]: E1013 08:45:13.502215 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:13 crc kubenswrapper[4685]: E1013 08:45:13.502327 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.559955 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.560017 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.560034 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.560057 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.560073 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:13Z","lastTransitionTime":"2025-10-13T08:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.662223 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.662272 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.662290 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.662314 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.662332 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:13Z","lastTransitionTime":"2025-10-13T08:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.764788 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.764845 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.764857 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.764885 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.764903 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:13Z","lastTransitionTime":"2025-10-13T08:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.866983 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.867031 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.867044 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.867061 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.867076 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:13Z","lastTransitionTime":"2025-10-13T08:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.969549 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.969596 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.969607 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.969625 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:13 crc kubenswrapper[4685]: I1013 08:45:13.969638 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:13Z","lastTransitionTime":"2025-10-13T08:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.071454 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.071494 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.071526 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.071542 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.071551 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:14Z","lastTransitionTime":"2025-10-13T08:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.173640 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.173694 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.173709 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.173726 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.173740 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:14Z","lastTransitionTime":"2025-10-13T08:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.276470 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.276534 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.276557 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.276584 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.276605 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:14Z","lastTransitionTime":"2025-10-13T08:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.378808 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.378857 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.378866 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.378882 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.378893 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:14Z","lastTransitionTime":"2025-10-13T08:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.481694 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.481772 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.481785 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.481803 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.481837 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:14Z","lastTransitionTime":"2025-10-13T08:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.502029 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.502353 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:14 crc kubenswrapper[4685]: E1013 08:45:14.505535 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:14 crc kubenswrapper[4685]: E1013 08:45:14.506697 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.575537 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs\") pod \"network-metrics-daemon-nkxgw\" (UID: \"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\") " pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:14 crc kubenswrapper[4685]: E1013 08:45:14.575716 4685 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 08:45:14 crc kubenswrapper[4685]: E1013 08:45:14.575832 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs podName:7cb40c32-d7d4-4a1b-9cda-1de0fab48566 nodeName:}" failed. No retries permitted until 2025-10-13 08:45:18.575808169 +0000 UTC m=+43.723684000 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs") pod "network-metrics-daemon-nkxgw" (UID: "7cb40c32-d7d4-4a1b-9cda-1de0fab48566") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.584310 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.584483 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.584572 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.584681 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.584767 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:14Z","lastTransitionTime":"2025-10-13T08:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.686526 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.686606 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.686619 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.686634 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.686644 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:14Z","lastTransitionTime":"2025-10-13T08:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.788848 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.789153 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.789226 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.789408 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.789549 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:14Z","lastTransitionTime":"2025-10-13T08:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.892580 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.892855 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.893202 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.893363 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.893492 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:14Z","lastTransitionTime":"2025-10-13T08:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.996406 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.996480 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.996502 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.996534 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:14 crc kubenswrapper[4685]: I1013 08:45:14.996561 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:14Z","lastTransitionTime":"2025-10-13T08:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.099810 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.099895 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.099921 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.099939 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.099951 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:15Z","lastTransitionTime":"2025-10-13T08:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.203464 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.203519 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.203532 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.203550 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.203562 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:15Z","lastTransitionTime":"2025-10-13T08:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.306509 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.306577 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.306599 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.306625 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.306642 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:15Z","lastTransitionTime":"2025-10-13T08:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.409328 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.409361 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.409371 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.409385 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.409393 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:15Z","lastTransitionTime":"2025-10-13T08:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.502139 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:15 crc kubenswrapper[4685]: E1013 08:45:15.502241 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.502354 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:15 crc kubenswrapper[4685]: E1013 08:45:15.502499 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.511745 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.511772 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.511781 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.511792 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.511817 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:15Z","lastTransitionTime":"2025-10-13T08:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.526761 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:15Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.541315 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:15Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.567020 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:15Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.612051 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:15Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.614244 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.614269 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.614278 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.614292 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.614301 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:15Z","lastTransitionTime":"2025-10-13T08:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.630370 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:15Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.643293 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:15Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.660177 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"message\\\":\\\"p:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.138:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97419c58-41c7-41d7-a137-a446f0c7eeb3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1013 08:45:08.975476 6040 factory.go:656] Stopping watch factory\\\\nI1013 08:45:08.975487 6040 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1013 08:45:08.975497 6040 ovnkube.go:599] Stopped ovnkube\\\\nI1013 08:45:08.975484 6040 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-operators]} name:Service_openshift-marketplace/redhat-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.138:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97419c58-41c7-41d7-a137-a446f0c7eeb3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1013 08:45:08.975531 6040 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1013 08:45:08.975648 6040 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5kgxk_openshift-ovn-kubernetes(ffa4857b-f741-47d8-b692-74a84be30411)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:15Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.672089 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:15Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.683781 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:15Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.692738 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:15Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.712273 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:15Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.715976 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.716005 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.716016 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.716031 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.716042 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:15Z","lastTransitionTime":"2025-10-13T08:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.723479 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:15Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.735895 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:15Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.746986 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:15Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.755941 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:15Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.765813 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295380a1-3f1f-4c5e-907c-b0545aa739e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf1b49a1ee1c45ddd3d90831ec5ea3064b05a7ac65ae67ccdfe3dc25ae0e5c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7136baf6f1e826c79c8171ce14480489ff2fab3c1918d6d807535624716b19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p8d8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:15Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.774714 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nkxgw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nkxgw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:15Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.818830 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.818869 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.818877 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.818892 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.818903 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:15Z","lastTransitionTime":"2025-10-13T08:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.921258 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.921303 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.921314 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.921331 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:15 crc kubenswrapper[4685]: I1013 08:45:15.921342 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:15Z","lastTransitionTime":"2025-10-13T08:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.024584 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.024633 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.024644 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.024663 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.024672 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:16Z","lastTransitionTime":"2025-10-13T08:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.127091 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.127148 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.127163 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.127184 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.127200 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:16Z","lastTransitionTime":"2025-10-13T08:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.229776 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.229834 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.229846 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.229865 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.229878 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:16Z","lastTransitionTime":"2025-10-13T08:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.332575 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.332621 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.332633 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.332651 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.332665 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:16Z","lastTransitionTime":"2025-10-13T08:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.434719 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.434768 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.434781 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.434800 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.434811 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:16Z","lastTransitionTime":"2025-10-13T08:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.502252 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.502252 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:16 crc kubenswrapper[4685]: E1013 08:45:16.502394 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:16 crc kubenswrapper[4685]: E1013 08:45:16.502436 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.537382 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.537613 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.537672 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.537747 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.537803 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:16Z","lastTransitionTime":"2025-10-13T08:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.639557 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.639594 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.639603 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.639616 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.639625 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:16Z","lastTransitionTime":"2025-10-13T08:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.742013 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.742376 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.742462 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.742575 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.742645 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:16Z","lastTransitionTime":"2025-10-13T08:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.845556 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.845591 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.845606 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.845620 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.845630 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:16Z","lastTransitionTime":"2025-10-13T08:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.948659 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.948707 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.948722 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.948768 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:16 crc kubenswrapper[4685]: I1013 08:45:16.948781 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:16Z","lastTransitionTime":"2025-10-13T08:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.050759 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.051401 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.051503 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.051586 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.051650 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:17Z","lastTransitionTime":"2025-10-13T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.153994 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.154030 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.154040 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.154056 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.154066 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:17Z","lastTransitionTime":"2025-10-13T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.256416 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.256450 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.256459 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.256473 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.256512 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:17Z","lastTransitionTime":"2025-10-13T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.358549 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.358957 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.359065 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.359186 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.359304 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:17Z","lastTransitionTime":"2025-10-13T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.466302 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.466344 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.466355 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.466370 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.466382 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:17Z","lastTransitionTime":"2025-10-13T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.501813 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.501833 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:17 crc kubenswrapper[4685]: E1013 08:45:17.502363 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:17 crc kubenswrapper[4685]: E1013 08:45:17.502301 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.568539 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.568767 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.568869 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.568957 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.569040 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:17Z","lastTransitionTime":"2025-10-13T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.671221 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.671250 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.671259 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.671273 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.671283 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:17Z","lastTransitionTime":"2025-10-13T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.774223 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.774497 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.774565 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.774652 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.774777 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:17Z","lastTransitionTime":"2025-10-13T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.876803 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.876837 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.876847 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.876861 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.876870 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:17Z","lastTransitionTime":"2025-10-13T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.979387 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.979436 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.979447 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.979465 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:17 crc kubenswrapper[4685]: I1013 08:45:17.979475 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:17Z","lastTransitionTime":"2025-10-13T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.081238 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.081281 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.081297 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.081318 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.081334 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:18Z","lastTransitionTime":"2025-10-13T08:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.183839 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.183894 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.183934 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.183959 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.183978 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:18Z","lastTransitionTime":"2025-10-13T08:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.287653 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.288002 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.288084 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.288160 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.288224 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:18Z","lastTransitionTime":"2025-10-13T08:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.391707 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.391746 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.391757 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.391773 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.391782 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:18Z","lastTransitionTime":"2025-10-13T08:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.493898 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.493945 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.493953 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.493967 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.493976 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:18Z","lastTransitionTime":"2025-10-13T08:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.502542 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:18 crc kubenswrapper[4685]: E1013 08:45:18.502630 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.502979 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:18 crc kubenswrapper[4685]: E1013 08:45:18.503203 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.596806 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.597140 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.597226 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.597299 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.597355 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:18Z","lastTransitionTime":"2025-10-13T08:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.615359 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs\") pod \"network-metrics-daemon-nkxgw\" (UID: \"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\") " pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:18 crc kubenswrapper[4685]: E1013 08:45:18.615585 4685 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 08:45:18 crc kubenswrapper[4685]: E1013 08:45:18.615703 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs podName:7cb40c32-d7d4-4a1b-9cda-1de0fab48566 nodeName:}" failed. No retries permitted until 2025-10-13 08:45:26.615687975 +0000 UTC m=+51.763563736 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs") pod "network-metrics-daemon-nkxgw" (UID: "7cb40c32-d7d4-4a1b-9cda-1de0fab48566") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.699442 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.699785 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.699871 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.700034 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.700153 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:18Z","lastTransitionTime":"2025-10-13T08:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.802482 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.802859 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.803140 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.803366 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.803543 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:18Z","lastTransitionTime":"2025-10-13T08:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.906357 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.906680 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.906888 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.907130 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:18 crc kubenswrapper[4685]: I1013 08:45:18.907409 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:18Z","lastTransitionTime":"2025-10-13T08:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.010463 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.010503 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.010513 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.010530 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.010540 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:19Z","lastTransitionTime":"2025-10-13T08:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.113088 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.113130 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.113139 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.113153 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.113165 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:19Z","lastTransitionTime":"2025-10-13T08:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.215793 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.215826 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.215836 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.215851 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.215862 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:19Z","lastTransitionTime":"2025-10-13T08:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.318243 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.318289 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.318301 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.318317 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.318328 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:19Z","lastTransitionTime":"2025-10-13T08:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.420722 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.420752 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.420760 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.420777 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.420785 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:19Z","lastTransitionTime":"2025-10-13T08:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.502602 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.502679 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:19 crc kubenswrapper[4685]: E1013 08:45:19.502726 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:19 crc kubenswrapper[4685]: E1013 08:45:19.502998 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.523416 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.523464 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.523474 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.523495 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.523506 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:19Z","lastTransitionTime":"2025-10-13T08:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.625796 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.625838 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.625849 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.625866 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.625881 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:19Z","lastTransitionTime":"2025-10-13T08:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.728623 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.728873 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.728961 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.729066 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.729130 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:19Z","lastTransitionTime":"2025-10-13T08:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.831974 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.832017 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.832026 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.832042 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.832052 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:19Z","lastTransitionTime":"2025-10-13T08:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.934601 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.934651 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.934666 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.934684 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:19 crc kubenswrapper[4685]: I1013 08:45:19.934697 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:19Z","lastTransitionTime":"2025-10-13T08:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.037399 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.037519 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.037538 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.037567 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.037586 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:20Z","lastTransitionTime":"2025-10-13T08:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.140133 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.140180 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.140191 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.140206 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.140216 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:20Z","lastTransitionTime":"2025-10-13T08:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.243784 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.244250 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.244386 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.244517 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.244630 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:20Z","lastTransitionTime":"2025-10-13T08:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.347882 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.347984 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.348000 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.348022 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.348034 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:20Z","lastTransitionTime":"2025-10-13T08:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.451176 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.451492 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.451618 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.451759 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.451962 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:20Z","lastTransitionTime":"2025-10-13T08:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.502020 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.502053 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:20 crc kubenswrapper[4685]: E1013 08:45:20.502283 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:20 crc kubenswrapper[4685]: E1013 08:45:20.502376 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.555080 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.555339 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.555430 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.555520 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.555612 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:20Z","lastTransitionTime":"2025-10-13T08:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.659171 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.659474 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.659553 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.659648 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.659727 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:20Z","lastTransitionTime":"2025-10-13T08:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.763398 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.763438 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.763448 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.763463 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.763474 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:20Z","lastTransitionTime":"2025-10-13T08:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.866459 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.866688 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.866777 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.866941 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.867011 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:20Z","lastTransitionTime":"2025-10-13T08:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.970074 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.970127 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.970137 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.970155 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:20 crc kubenswrapper[4685]: I1013 08:45:20.970168 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:20Z","lastTransitionTime":"2025-10-13T08:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.073150 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.073192 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.073203 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.073221 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.073233 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:21Z","lastTransitionTime":"2025-10-13T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.175452 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.175523 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.175546 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.175575 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.175597 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:21Z","lastTransitionTime":"2025-10-13T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.278183 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.278268 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.278300 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.278333 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.278356 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:21Z","lastTransitionTime":"2025-10-13T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.380572 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.380956 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.381118 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.381256 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.381346 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:21Z","lastTransitionTime":"2025-10-13T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.484256 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.484324 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.484335 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.484379 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.484397 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:21Z","lastTransitionTime":"2025-10-13T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.502691 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.502730 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:21 crc kubenswrapper[4685]: E1013 08:45:21.502821 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:21 crc kubenswrapper[4685]: E1013 08:45:21.502897 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.586697 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.586733 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.586743 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.586760 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.586771 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:21Z","lastTransitionTime":"2025-10-13T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.689604 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.689877 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.689979 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.690075 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.690145 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:21Z","lastTransitionTime":"2025-10-13T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.702751 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.702795 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.702811 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.702835 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.702851 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:21Z","lastTransitionTime":"2025-10-13T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:21 crc kubenswrapper[4685]: E1013 08:45:21.719877 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:21Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.725796 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.725961 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.726053 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.726154 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.726261 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:21Z","lastTransitionTime":"2025-10-13T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:21 crc kubenswrapper[4685]: E1013 08:45:21.746464 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:21Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.751594 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.751815 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.751935 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.752045 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.752137 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:21Z","lastTransitionTime":"2025-10-13T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:21 crc kubenswrapper[4685]: E1013 08:45:21.765361 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:21Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.769876 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.770042 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.770065 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.770139 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.770160 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:21Z","lastTransitionTime":"2025-10-13T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:21 crc kubenswrapper[4685]: E1013 08:45:21.790853 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:21Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.795426 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.795464 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.795474 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.795489 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.795502 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:21Z","lastTransitionTime":"2025-10-13T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:21 crc kubenswrapper[4685]: E1013 08:45:21.813053 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:21Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:21 crc kubenswrapper[4685]: E1013 08:45:21.813166 4685 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.820106 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.820193 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.820217 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.820249 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.820272 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:21Z","lastTransitionTime":"2025-10-13T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.923137 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.923172 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.923183 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.923197 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:21 crc kubenswrapper[4685]: I1013 08:45:21.923209 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:21Z","lastTransitionTime":"2025-10-13T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.026169 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.026568 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.026690 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.026810 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.026939 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:22Z","lastTransitionTime":"2025-10-13T08:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.130636 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.130728 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.130764 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.130792 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.130808 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:22Z","lastTransitionTime":"2025-10-13T08:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.234327 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.234475 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.234531 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.234565 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.234614 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:22Z","lastTransitionTime":"2025-10-13T08:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.338636 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.338723 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.338748 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.338786 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.338811 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:22Z","lastTransitionTime":"2025-10-13T08:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.441408 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.441671 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.441750 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.441825 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.441899 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:22Z","lastTransitionTime":"2025-10-13T08:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.502490 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.502519 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:22 crc kubenswrapper[4685]: E1013 08:45:22.502750 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:22 crc kubenswrapper[4685]: E1013 08:45:22.502832 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.545184 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.545223 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.545235 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.545253 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.545264 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:22Z","lastTransitionTime":"2025-10-13T08:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.647892 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.647986 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.648018 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.648037 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.648050 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:22Z","lastTransitionTime":"2025-10-13T08:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.750369 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.750408 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.750419 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.750433 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.750443 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:22Z","lastTransitionTime":"2025-10-13T08:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.853320 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.853380 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.853392 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.853413 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.853428 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:22Z","lastTransitionTime":"2025-10-13T08:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.955787 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.956124 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.956208 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.956288 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:22 crc kubenswrapper[4685]: I1013 08:45:22.956373 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:22Z","lastTransitionTime":"2025-10-13T08:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.058844 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.058904 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.058966 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.059001 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.059025 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:23Z","lastTransitionTime":"2025-10-13T08:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.161443 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.161503 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.161518 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.161539 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.161554 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:23Z","lastTransitionTime":"2025-10-13T08:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.263883 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.263939 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.263951 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.263968 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.263982 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:23Z","lastTransitionTime":"2025-10-13T08:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.366547 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.366597 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.366614 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.366634 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.366649 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:23Z","lastTransitionTime":"2025-10-13T08:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.469615 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.469661 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.469672 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.469691 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.469703 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:23Z","lastTransitionTime":"2025-10-13T08:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.502271 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:23 crc kubenswrapper[4685]: E1013 08:45:23.502491 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.502290 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:23 crc kubenswrapper[4685]: E1013 08:45:23.502644 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.503513 4685 scope.go:117] "RemoveContainer" containerID="4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.572308 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.572747 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.572767 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.573455 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.573482 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:23Z","lastTransitionTime":"2025-10-13T08:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.677173 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.677266 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.677279 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.677303 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.677315 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:23Z","lastTransitionTime":"2025-10-13T08:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.780084 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.780147 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.780159 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.780176 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.780189 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:23Z","lastTransitionTime":"2025-10-13T08:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.883806 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.883837 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.883845 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.883858 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.883866 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:23Z","lastTransitionTime":"2025-10-13T08:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.890812 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5kgxk_ffa4857b-f741-47d8-b692-74a84be30411/ovnkube-controller/1.log" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.894565 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerStarted","Data":"3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0"} Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.895046 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.911194 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:23Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.928510 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:23Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.943819 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295380a1-3f1f-4c5e-907c-b0545aa739e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf1b49a1ee1c45ddd3d90831ec5ea3064b05a7ac65ae67ccdfe3dc25ae0e5c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7136baf6f1e826c79c8171ce14480489ff2fab3c1918d6d807535624716b19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p8d8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:23Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.955848 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nkxgw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nkxgw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:23Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.982567 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:23Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.986990 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.987033 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.987044 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.987064 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:23 crc kubenswrapper[4685]: I1013 08:45:23.987080 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:23Z","lastTransitionTime":"2025-10-13T08:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.002942 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:23Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.020104 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:24Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.035963 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:24Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.052272 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:24Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.065689 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:24Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.079057 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:24Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.089327 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.089370 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.089379 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.089393 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.089402 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:24Z","lastTransitionTime":"2025-10-13T08:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.098638 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"message\\\":\\\"p:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.138:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97419c58-41c7-41d7-a137-a446f0c7eeb3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1013 08:45:08.975476 6040 factory.go:656] Stopping watch factory\\\\nI1013 08:45:08.975487 6040 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1013 08:45:08.975497 6040 ovnkube.go:599] Stopped ovnkube\\\\nI1013 08:45:08.975484 6040 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-operators]} name:Service_openshift-marketplace/redhat-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.138:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97419c58-41c7-41d7-a137-a446f0c7eeb3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1013 08:45:08.975531 6040 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1013 08:45:08.975648 6040 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:24Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.112486 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:24Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.125202 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:24Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.136579 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:24Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.153220 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:24Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.166375 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:24Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.192047 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.192119 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.192133 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.192151 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.192180 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:24Z","lastTransitionTime":"2025-10-13T08:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.294022 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.294074 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.294083 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.294098 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.294112 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:24Z","lastTransitionTime":"2025-10-13T08:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.396644 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.396703 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.396716 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.396732 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.396743 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:24Z","lastTransitionTime":"2025-10-13T08:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.499255 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.499297 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.499307 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.499323 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.499334 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:24Z","lastTransitionTime":"2025-10-13T08:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.502480 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.502485 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:24 crc kubenswrapper[4685]: E1013 08:45:24.502637 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:24 crc kubenswrapper[4685]: E1013 08:45:24.502731 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.602115 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.602154 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.602164 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.602179 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.602191 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:24Z","lastTransitionTime":"2025-10-13T08:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.704152 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.704195 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.704211 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.704233 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.704250 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:24Z","lastTransitionTime":"2025-10-13T08:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.806298 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.806329 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.806337 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.806350 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.806360 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:24Z","lastTransitionTime":"2025-10-13T08:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.900718 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5kgxk_ffa4857b-f741-47d8-b692-74a84be30411/ovnkube-controller/2.log" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.901510 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5kgxk_ffa4857b-f741-47d8-b692-74a84be30411/ovnkube-controller/1.log" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.904287 4685 generic.go:334] "Generic (PLEG): container finished" podID="ffa4857b-f741-47d8-b692-74a84be30411" containerID="3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0" exitCode=1 Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.904324 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerDied","Data":"3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0"} Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.904365 4685 scope.go:117] "RemoveContainer" containerID="4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.905701 4685 scope.go:117] "RemoveContainer" containerID="3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0" Oct 13 08:45:24 crc kubenswrapper[4685]: E1013 08:45:24.906052 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5kgxk_openshift-ovn-kubernetes(ffa4857b-f741-47d8-b692-74a84be30411)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" podUID="ffa4857b-f741-47d8-b692-74a84be30411" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.914802 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.914845 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.914859 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.914876 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.914886 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:24Z","lastTransitionTime":"2025-10-13T08:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.923394 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:24Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.935802 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:24Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.946343 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295380a1-3f1f-4c5e-907c-b0545aa739e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf1b49a1ee1c45ddd3d90831ec5ea3064b05a7ac65ae67ccdfe3dc25ae0e5c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7136baf6f1e826c79c8171ce14480489ff2fab3c1918d6d807535624716b19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p8d8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:24Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.956384 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nkxgw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nkxgw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:24Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.969949 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:24Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:24 crc kubenswrapper[4685]: I1013 08:45:24.983357 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:24Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.005052 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.017426 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.017649 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.017709 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.017803 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.017864 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:25Z","lastTransitionTime":"2025-10-13T08:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.022155 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.022185 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.034412 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.035441 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.055315 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"message\\\":\\\"p:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.138:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97419c58-41c7-41d7-a137-a446f0c7eeb3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1013 08:45:08.975476 6040 factory.go:656] Stopping watch factory\\\\nI1013 08:45:08.975487 6040 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1013 08:45:08.975497 6040 ovnkube.go:599] Stopped ovnkube\\\\nI1013 08:45:08.975484 6040 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-operators]} name:Service_openshift-marketplace/redhat-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.138:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97419c58-41c7-41d7-a137-a446f0c7eeb3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1013 08:45:08.975531 6040 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1013 08:45:08.975648 6040 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:24Z\\\",\\\"message\\\":\\\".Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1013 08:45:24.446602 6235 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager/kube-controller-manager for network=default are: map[]\\\\nI1013 08:45:24.448730 6235 services_controller.go:443] Built service openshift-kube-controller-manager/kube-controller-manager LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1013 08:45:24.448781 6235 services_controller.go:444] Built service openshift-kube-controller-manager/kube-controller-manager LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1013 08:45:24.448805 6235 services_controller.go:445] Built service openshift-kube-controller-manager/kube-controller-manager LB template configs for network=default: []services.lbConfig(nil)\\\\nF1013 08:45:24.448827 6235 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.071018 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.083030 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.094019 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.109124 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.118387 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.120539 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.120566 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.120576 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.120593 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.120604 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:25Z","lastTransitionTime":"2025-10-13T08:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.131152 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.142483 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.156899 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.170692 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.181961 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.196591 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.206525 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.214857 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nkxgw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nkxgw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.222850 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.222893 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.222904 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.222943 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.222954 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:25Z","lastTransitionTime":"2025-10-13T08:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.224903 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d37111fa-463a-422b-916d-c403ce10bf35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d110b9d78594b2f634d07714a305066c53a277b10a9d2ef513e47fc19f52b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://305186b83e0255ae886edebdb2038898352c016ab17ed4a594c7caac64babde0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16b91dfe5b7be866dc336876627a48fb7f5c3d4cfca022c4f7cf2cd448bb9510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.236394 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.245906 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.256365 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295380a1-3f1f-4c5e-907c-b0545aa739e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf1b49a1ee1c45ddd3d90831ec5ea3064b05a7ac65ae67ccdfe3dc25ae0e5c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7136baf6f1e826c79c8171ce14480489ff2fab3c1918d6d807535624716b19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p8d8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.273154 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.284631 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.296079 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.308476 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.321507 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.325048 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.325078 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.325086 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.325099 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.325108 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:25Z","lastTransitionTime":"2025-10-13T08:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.336686 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.347244 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.366514 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"message\\\":\\\"p:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.138:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97419c58-41c7-41d7-a137-a446f0c7eeb3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1013 08:45:08.975476 6040 factory.go:656] Stopping watch factory\\\\nI1013 08:45:08.975487 6040 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1013 08:45:08.975497 6040 ovnkube.go:599] Stopped ovnkube\\\\nI1013 08:45:08.975484 6040 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-operators]} name:Service_openshift-marketplace/redhat-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.138:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97419c58-41c7-41d7-a137-a446f0c7eeb3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1013 08:45:08.975531 6040 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1013 08:45:08.975648 6040 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:24Z\\\",\\\"message\\\":\\\".Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1013 08:45:24.446602 6235 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager/kube-controller-manager for network=default are: map[]\\\\nI1013 08:45:24.448730 6235 services_controller.go:443] Built service openshift-kube-controller-manager/kube-controller-manager LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1013 08:45:24.448781 6235 services_controller.go:444] Built service openshift-kube-controller-manager/kube-controller-manager LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1013 08:45:24.448805 6235 services_controller.go:445] Built service openshift-kube-controller-manager/kube-controller-manager LB template configs for network=default: []services.lbConfig(nil)\\\\nF1013 08:45:24.448827 6235 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.427618 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.428236 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.428406 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.428604 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.428759 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:25Z","lastTransitionTime":"2025-10-13T08:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.502621 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:25 crc kubenswrapper[4685]: E1013 08:45:25.502979 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.503326 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:25 crc kubenswrapper[4685]: E1013 08:45:25.503539 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.514371 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.527399 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.531760 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.531802 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.531817 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.531864 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.531875 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:25Z","lastTransitionTime":"2025-10-13T08:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.541054 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.553665 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.569260 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.582057 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295380a1-3f1f-4c5e-907c-b0545aa739e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf1b49a1ee1c45ddd3d90831ec5ea3064b05a7ac65ae67ccdfe3dc25ae0e5c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7136baf6f1e826c79c8171ce14480489ff2fab3c1918d6d807535624716b19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p8d8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.597161 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nkxgw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nkxgw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.610001 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d37111fa-463a-422b-916d-c403ce10bf35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d110b9d78594b2f634d07714a305066c53a277b10a9d2ef513e47fc19f52b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://305186b83e0255ae886edebdb2038898352c016ab17ed4a594c7caac64babde0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16b91dfe5b7be866dc336876627a48fb7f5c3d4cfca022c4f7cf2cd448bb9510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.622574 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.634198 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.634251 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.634260 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.634274 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.634282 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:25Z","lastTransitionTime":"2025-10-13T08:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.634663 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.653489 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.666031 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.677493 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.690247 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.702834 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.714903 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.727197 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.736598 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.736631 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.736641 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.736656 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.736667 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:25Z","lastTransitionTime":"2025-10-13T08:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.747755 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a8d39abe6c4457f8f6a0fccd3e3aa8a236a43ec1e7619072a4ee677b38db3fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"message\\\":\\\"p:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.138:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97419c58-41c7-41d7-a137-a446f0c7eeb3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1013 08:45:08.975476 6040 factory.go:656] Stopping watch factory\\\\nI1013 08:45:08.975487 6040 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1013 08:45:08.975497 6040 ovnkube.go:599] Stopped ovnkube\\\\nI1013 08:45:08.975484 6040 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/redhat-operators]} name:Service_openshift-marketplace/redhat-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.138:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97419c58-41c7-41d7-a137-a446f0c7eeb3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1013 08:45:08.975531 6040 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1013 08:45:08.975648 6040 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:24Z\\\",\\\"message\\\":\\\".Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1013 08:45:24.446602 6235 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager/kube-controller-manager for network=default are: map[]\\\\nI1013 08:45:24.448730 6235 services_controller.go:443] Built service openshift-kube-controller-manager/kube-controller-manager LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1013 08:45:24.448781 6235 services_controller.go:444] Built service openshift-kube-controller-manager/kube-controller-manager LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1013 08:45:24.448805 6235 services_controller.go:445] Built service openshift-kube-controller-manager/kube-controller-manager LB template configs for network=default: []services.lbConfig(nil)\\\\nF1013 08:45:24.448827 6235 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.838464 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.838499 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.838507 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.838521 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.838530 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:25Z","lastTransitionTime":"2025-10-13T08:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.908668 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5kgxk_ffa4857b-f741-47d8-b692-74a84be30411/ovnkube-controller/2.log" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.912201 4685 scope.go:117] "RemoveContainer" containerID="3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0" Oct 13 08:45:25 crc kubenswrapper[4685]: E1013 08:45:25.912363 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5kgxk_openshift-ovn-kubernetes(ffa4857b-f741-47d8-b692-74a84be30411)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" podUID="ffa4857b-f741-47d8-b692-74a84be30411" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.921204 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.931735 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.941021 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.941055 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.941065 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.941083 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.941095 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:25Z","lastTransitionTime":"2025-10-13T08:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.942412 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.952248 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.966864 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.977441 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295380a1-3f1f-4c5e-907c-b0545aa739e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf1b49a1ee1c45ddd3d90831ec5ea3064b05a7ac65ae67ccdfe3dc25ae0e5c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7136baf6f1e826c79c8171ce14480489ff2fab3c1918d6d807535624716b19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p8d8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.986534 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nkxgw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nkxgw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:25 crc kubenswrapper[4685]: I1013 08:45:25.997345 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d37111fa-463a-422b-916d-c403ce10bf35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d110b9d78594b2f634d07714a305066c53a277b10a9d2ef513e47fc19f52b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://305186b83e0255ae886edebdb2038898352c016ab17ed4a594c7caac64babde0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16b91dfe5b7be866dc336876627a48fb7f5c3d4cfca022c4f7cf2cd448bb9510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:25Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.009951 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:26Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.019771 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:26Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.036652 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:26Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.044514 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.044558 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.044567 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.044586 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.044595 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:26Z","lastTransitionTime":"2025-10-13T08:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.050848 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:26Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.063706 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:26Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.074749 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:26Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.086563 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:26Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.097634 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:26Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.113976 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:26Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.130273 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:24Z\\\",\\\"message\\\":\\\".Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1013 08:45:24.446602 6235 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager/kube-controller-manager for network=default are: map[]\\\\nI1013 08:45:24.448730 6235 services_controller.go:443] Built service openshift-kube-controller-manager/kube-controller-manager LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1013 08:45:24.448781 6235 services_controller.go:444] Built service openshift-kube-controller-manager/kube-controller-manager LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1013 08:45:24.448805 6235 services_controller.go:445] Built service openshift-kube-controller-manager/kube-controller-manager LB template configs for network=default: []services.lbConfig(nil)\\\\nF1013 08:45:24.448827 6235 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5kgxk_openshift-ovn-kubernetes(ffa4857b-f741-47d8-b692-74a84be30411)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:26Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.147297 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.147366 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.147378 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.147396 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.147408 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:26Z","lastTransitionTime":"2025-10-13T08:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.249825 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.249892 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.249956 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.249989 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.250050 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:26Z","lastTransitionTime":"2025-10-13T08:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.352891 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.352946 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.352958 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.352976 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.352988 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:26Z","lastTransitionTime":"2025-10-13T08:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.455621 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.455654 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.455662 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.455676 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.455685 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:26Z","lastTransitionTime":"2025-10-13T08:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.501639 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.501703 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:26 crc kubenswrapper[4685]: E1013 08:45:26.501819 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:26 crc kubenswrapper[4685]: E1013 08:45:26.501982 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.558708 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.559002 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.559106 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.559179 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.559236 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:26Z","lastTransitionTime":"2025-10-13T08:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.662639 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.662723 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.662747 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.662777 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.662799 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:26Z","lastTransitionTime":"2025-10-13T08:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.697571 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs\") pod \"network-metrics-daemon-nkxgw\" (UID: \"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\") " pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:26 crc kubenswrapper[4685]: E1013 08:45:26.697872 4685 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 08:45:26 crc kubenswrapper[4685]: E1013 08:45:26.698054 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs podName:7cb40c32-d7d4-4a1b-9cda-1de0fab48566 nodeName:}" failed. No retries permitted until 2025-10-13 08:45:42.698017515 +0000 UTC m=+67.845893316 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs") pod "network-metrics-daemon-nkxgw" (UID: "7cb40c32-d7d4-4a1b-9cda-1de0fab48566") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.765979 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.766293 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.766382 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.766471 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.766563 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:26Z","lastTransitionTime":"2025-10-13T08:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.869019 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.869336 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.869470 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.869606 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.869725 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:26Z","lastTransitionTime":"2025-10-13T08:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.973378 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.973765 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.973900 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.974073 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:26 crc kubenswrapper[4685]: I1013 08:45:26.974205 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:26Z","lastTransitionTime":"2025-10-13T08:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.078131 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.078200 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.078219 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.078244 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.078263 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:27Z","lastTransitionTime":"2025-10-13T08:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.182038 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.182073 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.182083 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.182100 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.182112 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:27Z","lastTransitionTime":"2025-10-13T08:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.202671 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.202854 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.202948 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:27 crc kubenswrapper[4685]: E1013 08:45:27.203102 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:45:59.203083029 +0000 UTC m=+84.350958800 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:45:27 crc kubenswrapper[4685]: E1013 08:45:27.203120 4685 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 08:45:27 crc kubenswrapper[4685]: E1013 08:45:27.203157 4685 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 08:45:27 crc kubenswrapper[4685]: E1013 08:45:27.203182 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 08:45:59.203164621 +0000 UTC m=+84.351040422 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 08:45:27 crc kubenswrapper[4685]: E1013 08:45:27.203212 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 08:45:59.203197702 +0000 UTC m=+84.351073503 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.285124 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.285172 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.285187 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.285208 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.285225 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:27Z","lastTransitionTime":"2025-10-13T08:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.304340 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.304403 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:27 crc kubenswrapper[4685]: E1013 08:45:27.304492 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 08:45:27 crc kubenswrapper[4685]: E1013 08:45:27.304522 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 08:45:27 crc kubenswrapper[4685]: E1013 08:45:27.304541 4685 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:45:27 crc kubenswrapper[4685]: E1013 08:45:27.304587 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 08:45:27 crc kubenswrapper[4685]: E1013 08:45:27.304607 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 08:45:27 crc kubenswrapper[4685]: E1013 08:45:27.304621 4685 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:45:27 crc kubenswrapper[4685]: E1013 08:45:27.304608 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-13 08:45:59.30458743 +0000 UTC m=+84.452463201 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:45:27 crc kubenswrapper[4685]: E1013 08:45:27.304673 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-13 08:45:59.304659492 +0000 UTC m=+84.452535263 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.387948 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.387989 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.388000 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.388014 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.388023 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:27Z","lastTransitionTime":"2025-10-13T08:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.490234 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.490301 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.490314 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.490330 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.490751 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:27Z","lastTransitionTime":"2025-10-13T08:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.502508 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.502636 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:27 crc kubenswrapper[4685]: E1013 08:45:27.502660 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:27 crc kubenswrapper[4685]: E1013 08:45:27.502806 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.593726 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.593770 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.593784 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.593806 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.593821 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:27Z","lastTransitionTime":"2025-10-13T08:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.696458 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.696513 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.696526 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.696546 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.696558 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:27Z","lastTransitionTime":"2025-10-13T08:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.799014 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.799072 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.799083 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.799104 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.799115 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:27Z","lastTransitionTime":"2025-10-13T08:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.901352 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.901686 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.901828 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.902105 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:27 crc kubenswrapper[4685]: I1013 08:45:27.902253 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:27Z","lastTransitionTime":"2025-10-13T08:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.004429 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.004722 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.004796 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.004867 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.004966 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:28Z","lastTransitionTime":"2025-10-13T08:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.107631 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.108053 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.108071 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.108091 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.108104 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:28Z","lastTransitionTime":"2025-10-13T08:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.210595 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.210664 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.210678 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.210699 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.210715 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:28Z","lastTransitionTime":"2025-10-13T08:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.313576 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.313657 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.313671 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.313689 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.313699 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:28Z","lastTransitionTime":"2025-10-13T08:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.417337 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.417386 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.417396 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.417413 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.417433 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:28Z","lastTransitionTime":"2025-10-13T08:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.502217 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:28 crc kubenswrapper[4685]: E1013 08:45:28.502357 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.502720 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:28 crc kubenswrapper[4685]: E1013 08:45:28.502792 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.520458 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.520515 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.520533 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.520563 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.520583 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:28Z","lastTransitionTime":"2025-10-13T08:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.623353 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.623381 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.623389 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.623403 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.623414 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:28Z","lastTransitionTime":"2025-10-13T08:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.726050 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.726089 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.726103 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.726123 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.726138 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:28Z","lastTransitionTime":"2025-10-13T08:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.828628 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.828662 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.828671 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.828687 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.828700 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:28Z","lastTransitionTime":"2025-10-13T08:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.931815 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.931876 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.931895 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.931959 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:28 crc kubenswrapper[4685]: I1013 08:45:28.931991 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:28Z","lastTransitionTime":"2025-10-13T08:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.035067 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.035123 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.035143 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.035169 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.035189 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:29Z","lastTransitionTime":"2025-10-13T08:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.138607 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.138658 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.138676 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.138704 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.138726 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:29Z","lastTransitionTime":"2025-10-13T08:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.241620 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.241976 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.242132 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.242262 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.242384 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:29Z","lastTransitionTime":"2025-10-13T08:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.345868 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.345957 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.345976 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.346001 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.346021 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:29Z","lastTransitionTime":"2025-10-13T08:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.449587 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.449653 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.449671 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.449697 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.449717 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:29Z","lastTransitionTime":"2025-10-13T08:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.502823 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.502823 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:29 crc kubenswrapper[4685]: E1013 08:45:29.503047 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:29 crc kubenswrapper[4685]: E1013 08:45:29.503203 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.553379 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.553451 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.553477 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.553513 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.553537 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:29Z","lastTransitionTime":"2025-10-13T08:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.657249 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.657339 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.657360 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.657389 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.657408 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:29Z","lastTransitionTime":"2025-10-13T08:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.760288 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.760735 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.760816 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.760898 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.761013 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:29Z","lastTransitionTime":"2025-10-13T08:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.864235 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.864290 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.864304 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.864321 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.864334 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:29Z","lastTransitionTime":"2025-10-13T08:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.966652 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.966682 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.966691 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.966706 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:29 crc kubenswrapper[4685]: I1013 08:45:29.966715 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:29Z","lastTransitionTime":"2025-10-13T08:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.069635 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.069678 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.069689 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.069706 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.069717 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:30Z","lastTransitionTime":"2025-10-13T08:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.172137 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.172167 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.172177 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.172192 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.172200 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:30Z","lastTransitionTime":"2025-10-13T08:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.274318 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.274364 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.274373 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.274397 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.274407 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:30Z","lastTransitionTime":"2025-10-13T08:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.376874 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.376905 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.376928 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.376970 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.376979 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:30Z","lastTransitionTime":"2025-10-13T08:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.479636 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.480282 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.480427 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.480530 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.480622 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:30Z","lastTransitionTime":"2025-10-13T08:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.501887 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.501946 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:30 crc kubenswrapper[4685]: E1013 08:45:30.502073 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:30 crc kubenswrapper[4685]: E1013 08:45:30.502153 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.583011 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.583044 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.583052 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.583065 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.583075 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:30Z","lastTransitionTime":"2025-10-13T08:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.685497 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.685533 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.685543 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.685559 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.685570 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:30Z","lastTransitionTime":"2025-10-13T08:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.791453 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.791967 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.792074 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.792257 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.792336 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:30Z","lastTransitionTime":"2025-10-13T08:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.895239 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.895291 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.895306 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.895328 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.895344 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:30Z","lastTransitionTime":"2025-10-13T08:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.998358 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.998390 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.998398 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.998413 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:30 crc kubenswrapper[4685]: I1013 08:45:30.998422 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:30Z","lastTransitionTime":"2025-10-13T08:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.101165 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.101195 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.101203 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.101218 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.101229 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:31Z","lastTransitionTime":"2025-10-13T08:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.203524 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.203572 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.203589 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.203612 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.203630 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:31Z","lastTransitionTime":"2025-10-13T08:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.306533 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.306745 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.306761 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.306780 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.306793 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:31Z","lastTransitionTime":"2025-10-13T08:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.409571 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.409634 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.409643 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.409660 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.409669 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:31Z","lastTransitionTime":"2025-10-13T08:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.502717 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.502820 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:31 crc kubenswrapper[4685]: E1013 08:45:31.502990 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:31 crc kubenswrapper[4685]: E1013 08:45:31.503164 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.511490 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.511537 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.511549 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.511568 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.511604 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:31Z","lastTransitionTime":"2025-10-13T08:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.613866 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.613944 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.613963 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.613979 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.613989 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:31Z","lastTransitionTime":"2025-10-13T08:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.716190 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.716231 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.716243 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.716261 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.716273 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:31Z","lastTransitionTime":"2025-10-13T08:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.819095 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.819146 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.819160 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.819177 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.819189 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:31Z","lastTransitionTime":"2025-10-13T08:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.921955 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.922003 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.922013 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.922030 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:31 crc kubenswrapper[4685]: I1013 08:45:31.922042 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:31Z","lastTransitionTime":"2025-10-13T08:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.024120 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.024171 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.024182 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.024197 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.024207 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:32Z","lastTransitionTime":"2025-10-13T08:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.112248 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.112302 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.112319 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.112343 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.112363 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:32Z","lastTransitionTime":"2025-10-13T08:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:32 crc kubenswrapper[4685]: E1013 08:45:32.124426 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:32Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.127749 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.127783 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.127812 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.127832 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.127843 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:32Z","lastTransitionTime":"2025-10-13T08:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:32 crc kubenswrapper[4685]: E1013 08:45:32.139061 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:32Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.142605 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.142632 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.142640 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.142653 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.142662 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:32Z","lastTransitionTime":"2025-10-13T08:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:32 crc kubenswrapper[4685]: E1013 08:45:32.156774 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:32Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.160701 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.160725 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.160733 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.160746 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.160755 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:32Z","lastTransitionTime":"2025-10-13T08:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:32 crc kubenswrapper[4685]: E1013 08:45:32.173518 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:32Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.177659 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.177713 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.177723 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.177736 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.177745 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:32Z","lastTransitionTime":"2025-10-13T08:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:32 crc kubenswrapper[4685]: E1013 08:45:32.191879 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:32Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:32 crc kubenswrapper[4685]: E1013 08:45:32.192033 4685 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.193511 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.193539 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.193550 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.193566 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.193581 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:32Z","lastTransitionTime":"2025-10-13T08:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.295713 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.296074 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.296257 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.296284 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.296297 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:32Z","lastTransitionTime":"2025-10-13T08:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.398576 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.398871 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.398990 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.399074 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.399150 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:32Z","lastTransitionTime":"2025-10-13T08:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.501429 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.501461 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.501472 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.501487 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.501499 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:32Z","lastTransitionTime":"2025-10-13T08:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.501723 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:32 crc kubenswrapper[4685]: E1013 08:45:32.501814 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.501877 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:32 crc kubenswrapper[4685]: E1013 08:45:32.501967 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.605035 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.605072 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.605081 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.605095 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.605104 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:32Z","lastTransitionTime":"2025-10-13T08:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.707893 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.708258 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.708335 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.708423 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.708544 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:32Z","lastTransitionTime":"2025-10-13T08:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.811660 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.811740 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.811755 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.811779 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.811792 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:32Z","lastTransitionTime":"2025-10-13T08:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.914173 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.914231 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.914246 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.914269 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:32 crc kubenswrapper[4685]: I1013 08:45:32.914287 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:32Z","lastTransitionTime":"2025-10-13T08:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.016491 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.016567 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.016577 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.016595 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.016605 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:33Z","lastTransitionTime":"2025-10-13T08:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.119651 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.119723 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.119732 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.119749 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.119759 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:33Z","lastTransitionTime":"2025-10-13T08:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.222515 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.222571 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.222581 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.222606 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.222619 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:33Z","lastTransitionTime":"2025-10-13T08:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.325060 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.325141 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.325155 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.325210 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.325229 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:33Z","lastTransitionTime":"2025-10-13T08:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.428139 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.428185 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.428197 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.428214 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.428227 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:33Z","lastTransitionTime":"2025-10-13T08:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.502483 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.502491 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:33 crc kubenswrapper[4685]: E1013 08:45:33.502790 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:33 crc kubenswrapper[4685]: E1013 08:45:33.502941 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.530518 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.530564 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.530574 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.530595 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.530608 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:33Z","lastTransitionTime":"2025-10-13T08:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.633415 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.633454 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.633463 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.633477 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.633486 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:33Z","lastTransitionTime":"2025-10-13T08:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.736752 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.736799 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.736810 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.736827 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.736843 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:33Z","lastTransitionTime":"2025-10-13T08:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.839859 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.839930 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.839943 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.839959 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.839968 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:33Z","lastTransitionTime":"2025-10-13T08:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.942364 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.942405 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.942417 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.942436 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:33 crc kubenswrapper[4685]: I1013 08:45:33.942447 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:33Z","lastTransitionTime":"2025-10-13T08:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.044712 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.044748 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.044758 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.044773 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.044782 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:34Z","lastTransitionTime":"2025-10-13T08:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.147402 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.147436 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.147453 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.147468 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.147478 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:34Z","lastTransitionTime":"2025-10-13T08:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.250170 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.250247 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.250262 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.250290 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.250307 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:34Z","lastTransitionTime":"2025-10-13T08:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.353032 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.353076 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.353089 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.353109 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.353123 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:34Z","lastTransitionTime":"2025-10-13T08:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.456983 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.457087 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.457107 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.457138 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.457158 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:34Z","lastTransitionTime":"2025-10-13T08:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.502039 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.502207 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:34 crc kubenswrapper[4685]: E1013 08:45:34.502261 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:34 crc kubenswrapper[4685]: E1013 08:45:34.502411 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.561683 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.561745 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.561759 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.561780 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.561793 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:34Z","lastTransitionTime":"2025-10-13T08:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.665156 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.665206 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.665219 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.665240 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.665253 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:34Z","lastTransitionTime":"2025-10-13T08:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.768359 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.768402 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.768413 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.768431 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.768444 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:34Z","lastTransitionTime":"2025-10-13T08:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.870879 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.870922 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.870932 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.870947 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.870956 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:34Z","lastTransitionTime":"2025-10-13T08:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.974029 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.974366 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.974539 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.974975 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:34 crc kubenswrapper[4685]: I1013 08:45:34.975169 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:34Z","lastTransitionTime":"2025-10-13T08:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.078440 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.078491 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.078504 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.078526 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.078539 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:35Z","lastTransitionTime":"2025-10-13T08:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.181574 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.181616 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.181625 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.181640 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.181650 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:35Z","lastTransitionTime":"2025-10-13T08:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.286966 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.287002 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.287014 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.287030 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.287045 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:35Z","lastTransitionTime":"2025-10-13T08:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.389533 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.389586 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.389596 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.389613 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.389625 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:35Z","lastTransitionTime":"2025-10-13T08:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.492361 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.492425 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.492435 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.492478 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.492500 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:35Z","lastTransitionTime":"2025-10-13T08:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.502567 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:35 crc kubenswrapper[4685]: E1013 08:45:35.502791 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.502599 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:35 crc kubenswrapper[4685]: E1013 08:45:35.503043 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.537296 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:35Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.555694 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:35Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.572227 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:35Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.590029 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:35Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.594600 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.594737 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.594775 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.594807 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.594818 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:35Z","lastTransitionTime":"2025-10-13T08:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.612576 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:35Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.629049 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:35Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.643797 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:35Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.665493 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:24Z\\\",\\\"message\\\":\\\".Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1013 08:45:24.446602 6235 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager/kube-controller-manager for network=default are: map[]\\\\nI1013 08:45:24.448730 6235 services_controller.go:443] Built service openshift-kube-controller-manager/kube-controller-manager LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1013 08:45:24.448781 6235 services_controller.go:444] Built service openshift-kube-controller-manager/kube-controller-manager LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1013 08:45:24.448805 6235 services_controller.go:445] Built service openshift-kube-controller-manager/kube-controller-manager LB template configs for network=default: []services.lbConfig(nil)\\\\nF1013 08:45:24.448827 6235 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5kgxk_openshift-ovn-kubernetes(ffa4857b-f741-47d8-b692-74a84be30411)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:35Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.680071 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:35Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.692771 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:35Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.697537 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.697574 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.697584 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.697603 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.697614 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:35Z","lastTransitionTime":"2025-10-13T08:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.705988 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:35Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.721251 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:35Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.733691 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:35Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.747132 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d37111fa-463a-422b-916d-c403ce10bf35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d110b9d78594b2f634d07714a305066c53a277b10a9d2ef513e47fc19f52b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://305186b83e0255ae886edebdb2038898352c016ab17ed4a594c7caac64babde0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16b91dfe5b7be866dc336876627a48fb7f5c3d4cfca022c4f7cf2cd448bb9510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:35Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.761936 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:35Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.780832 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:35Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.799579 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295380a1-3f1f-4c5e-907c-b0545aa739e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf1b49a1ee1c45ddd3d90831ec5ea3064b05a7ac65ae67ccdfe3dc25ae0e5c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7136baf6f1e826c79c8171ce14480489ff2fab3c1918d6d807535624716b19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p8d8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:35Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.800661 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.800697 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.800709 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.800737 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.800756 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:35Z","lastTransitionTime":"2025-10-13T08:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.815602 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nkxgw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nkxgw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:35Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.904799 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.904836 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.904847 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.904867 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:35 crc kubenswrapper[4685]: I1013 08:45:35.904879 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:35Z","lastTransitionTime":"2025-10-13T08:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.008373 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.008443 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.008461 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.008488 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.008509 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:36Z","lastTransitionTime":"2025-10-13T08:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.112779 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.112841 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.112855 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.112881 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.112895 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:36Z","lastTransitionTime":"2025-10-13T08:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.216551 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.216597 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.216610 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.216627 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.216640 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:36Z","lastTransitionTime":"2025-10-13T08:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.319316 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.319372 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.319385 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.319405 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.319419 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:36Z","lastTransitionTime":"2025-10-13T08:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.422714 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.422785 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.422801 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.422830 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.422849 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:36Z","lastTransitionTime":"2025-10-13T08:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.502505 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.502612 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:36 crc kubenswrapper[4685]: E1013 08:45:36.502674 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:36 crc kubenswrapper[4685]: E1013 08:45:36.502794 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.525076 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.525154 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.525172 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.525201 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.525220 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:36Z","lastTransitionTime":"2025-10-13T08:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.627905 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.627965 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.627979 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.627997 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.628008 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:36Z","lastTransitionTime":"2025-10-13T08:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.730565 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.730619 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.730639 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.730663 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.730678 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:36Z","lastTransitionTime":"2025-10-13T08:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.832962 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.833023 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.833033 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.833061 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.833076 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:36Z","lastTransitionTime":"2025-10-13T08:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.936524 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.936573 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.936587 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.936606 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:36 crc kubenswrapper[4685]: I1013 08:45:36.936619 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:36Z","lastTransitionTime":"2025-10-13T08:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.040073 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.040122 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.040132 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.040151 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.040162 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:37Z","lastTransitionTime":"2025-10-13T08:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.143628 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.143660 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.143668 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.143683 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.143692 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:37Z","lastTransitionTime":"2025-10-13T08:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.247774 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.247842 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.247858 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.247878 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.247893 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:37Z","lastTransitionTime":"2025-10-13T08:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.352398 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.352449 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.352460 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.352484 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.352498 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:37Z","lastTransitionTime":"2025-10-13T08:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.456860 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.456989 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.457017 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.457058 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.457120 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:37Z","lastTransitionTime":"2025-10-13T08:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.502610 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:37 crc kubenswrapper[4685]: E1013 08:45:37.502856 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.503442 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:37 crc kubenswrapper[4685]: E1013 08:45:37.503595 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.560559 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.560630 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.560647 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.560708 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.560728 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:37Z","lastTransitionTime":"2025-10-13T08:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.663746 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.663830 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.663854 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.663889 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.663946 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:37Z","lastTransitionTime":"2025-10-13T08:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.766996 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.767121 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.767141 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.767166 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.767181 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:37Z","lastTransitionTime":"2025-10-13T08:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.869980 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.870026 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.870039 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.870056 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.870069 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:37Z","lastTransitionTime":"2025-10-13T08:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.972358 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.972391 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.972399 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.972414 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:37 crc kubenswrapper[4685]: I1013 08:45:37.972426 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:37Z","lastTransitionTime":"2025-10-13T08:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.074616 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.074661 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.074671 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.074688 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.074698 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:38Z","lastTransitionTime":"2025-10-13T08:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.177405 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.177445 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.177454 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.177470 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.177483 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:38Z","lastTransitionTime":"2025-10-13T08:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.280349 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.280383 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.280394 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.280411 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.280424 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:38Z","lastTransitionTime":"2025-10-13T08:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.382790 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.382825 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.382834 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.382852 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.382863 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:38Z","lastTransitionTime":"2025-10-13T08:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.488655 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.488690 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.488698 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.488712 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.488721 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:38Z","lastTransitionTime":"2025-10-13T08:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.502093 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:38 crc kubenswrapper[4685]: E1013 08:45:38.502191 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.502197 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:38 crc kubenswrapper[4685]: E1013 08:45:38.502430 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.591696 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.591748 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.591763 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.591781 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.591793 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:38Z","lastTransitionTime":"2025-10-13T08:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.694635 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.694680 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.694692 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.694708 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.694718 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:38Z","lastTransitionTime":"2025-10-13T08:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.798020 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.798101 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.798112 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.798129 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.798140 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:38Z","lastTransitionTime":"2025-10-13T08:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.901557 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.903032 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.903069 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.903090 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:38 crc kubenswrapper[4685]: I1013 08:45:38.903103 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:38Z","lastTransitionTime":"2025-10-13T08:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.005641 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.005672 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.005681 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.005694 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.005702 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:39Z","lastTransitionTime":"2025-10-13T08:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.107558 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.107606 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.107617 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.107633 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.107643 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:39Z","lastTransitionTime":"2025-10-13T08:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.210763 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.211133 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.211207 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.211297 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.211405 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:39Z","lastTransitionTime":"2025-10-13T08:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.314560 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.314837 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.314943 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.315024 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.315097 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:39Z","lastTransitionTime":"2025-10-13T08:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.418441 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.419737 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.419853 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.419940 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.420001 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:39Z","lastTransitionTime":"2025-10-13T08:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.502334 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:39 crc kubenswrapper[4685]: E1013 08:45:39.502575 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.502952 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:39 crc kubenswrapper[4685]: E1013 08:45:39.503060 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.523296 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.523360 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.523381 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.523411 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.523433 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:39Z","lastTransitionTime":"2025-10-13T08:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.625882 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.626286 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.626611 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.626718 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.626795 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:39Z","lastTransitionTime":"2025-10-13T08:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.734454 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.734494 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.734503 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.734519 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.734529 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:39Z","lastTransitionTime":"2025-10-13T08:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.836672 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.837375 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.837473 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.837568 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.837658 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:39Z","lastTransitionTime":"2025-10-13T08:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.940576 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.940854 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.941001 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.941095 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:39 crc kubenswrapper[4685]: I1013 08:45:39.941181 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:39Z","lastTransitionTime":"2025-10-13T08:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.045340 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.045387 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.045403 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.045419 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.045428 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:40Z","lastTransitionTime":"2025-10-13T08:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.147636 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.147677 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.147689 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.147707 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.147719 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:40Z","lastTransitionTime":"2025-10-13T08:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.250345 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.251170 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.251276 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.251370 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.251460 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:40Z","lastTransitionTime":"2025-10-13T08:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.366468 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.366519 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.366530 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.366547 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.366557 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:40Z","lastTransitionTime":"2025-10-13T08:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.469142 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.469416 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.469517 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.469639 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.469726 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:40Z","lastTransitionTime":"2025-10-13T08:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.501729 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.502277 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:40 crc kubenswrapper[4685]: E1013 08:45:40.502399 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:40 crc kubenswrapper[4685]: E1013 08:45:40.502823 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.503050 4685 scope.go:117] "RemoveContainer" containerID="3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0" Oct 13 08:45:40 crc kubenswrapper[4685]: E1013 08:45:40.503280 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5kgxk_openshift-ovn-kubernetes(ffa4857b-f741-47d8-b692-74a84be30411)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" podUID="ffa4857b-f741-47d8-b692-74a84be30411" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.572682 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.572730 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.572743 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.572759 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.572769 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:40Z","lastTransitionTime":"2025-10-13T08:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.674797 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.674836 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.674847 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.674863 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.674876 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:40Z","lastTransitionTime":"2025-10-13T08:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.778396 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.778432 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.778445 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.778463 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.778478 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:40Z","lastTransitionTime":"2025-10-13T08:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.880370 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.880424 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.880435 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.880450 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.880460 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:40Z","lastTransitionTime":"2025-10-13T08:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.983211 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.983247 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.983263 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.983348 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:40 crc kubenswrapper[4685]: I1013 08:45:40.983363 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:40Z","lastTransitionTime":"2025-10-13T08:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.085110 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.085138 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.085146 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.085158 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.085167 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:41Z","lastTransitionTime":"2025-10-13T08:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.187215 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.187269 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.187278 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.187293 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.187303 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:41Z","lastTransitionTime":"2025-10-13T08:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.288959 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.289000 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.289010 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.289024 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.289034 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:41Z","lastTransitionTime":"2025-10-13T08:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.391431 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.391756 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.391853 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.391974 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.392068 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:41Z","lastTransitionTime":"2025-10-13T08:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.495809 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.495941 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.495957 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.495982 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.496001 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:41Z","lastTransitionTime":"2025-10-13T08:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.501853 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.501973 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:41 crc kubenswrapper[4685]: E1013 08:45:41.502185 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:41 crc kubenswrapper[4685]: E1013 08:45:41.502223 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.598415 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.598680 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.598782 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.598854 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.598942 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:41Z","lastTransitionTime":"2025-10-13T08:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.702306 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.702356 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.702366 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.702385 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.702396 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:41Z","lastTransitionTime":"2025-10-13T08:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.804408 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.804678 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.804751 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.804826 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.804892 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:41Z","lastTransitionTime":"2025-10-13T08:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.907248 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.907787 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.907869 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.907994 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:41 crc kubenswrapper[4685]: I1013 08:45:41.908071 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:41Z","lastTransitionTime":"2025-10-13T08:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.010880 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.011199 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.011270 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.011331 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.011390 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:42Z","lastTransitionTime":"2025-10-13T08:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.113148 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.113188 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.113196 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.113211 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.113220 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:42Z","lastTransitionTime":"2025-10-13T08:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.216117 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.216165 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.216178 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.216198 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.216211 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:42Z","lastTransitionTime":"2025-10-13T08:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.318617 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.318673 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.318685 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.318706 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.318718 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:42Z","lastTransitionTime":"2025-10-13T08:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.421591 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.421643 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.421655 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.421673 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.421685 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:42Z","lastTransitionTime":"2025-10-13T08:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.428259 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.428298 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.428310 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.428327 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.428340 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:42Z","lastTransitionTime":"2025-10-13T08:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:42 crc kubenswrapper[4685]: E1013 08:45:42.441795 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:42Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.445832 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.445876 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.445890 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.445923 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.445960 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:42Z","lastTransitionTime":"2025-10-13T08:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:42 crc kubenswrapper[4685]: E1013 08:45:42.458329 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:42Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.461630 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.461706 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.461721 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.461740 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.461754 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:42Z","lastTransitionTime":"2025-10-13T08:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:42 crc kubenswrapper[4685]: E1013 08:45:42.475403 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:42Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.479326 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.479379 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.479393 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.479411 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.479442 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:42Z","lastTransitionTime":"2025-10-13T08:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:42 crc kubenswrapper[4685]: E1013 08:45:42.490853 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:42Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.494015 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.494049 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.494062 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.494081 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.494092 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:42Z","lastTransitionTime":"2025-10-13T08:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.502568 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.502628 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:42 crc kubenswrapper[4685]: E1013 08:45:42.502886 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:42 crc kubenswrapper[4685]: E1013 08:45:42.502978 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:42 crc kubenswrapper[4685]: E1013 08:45:42.506201 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:42Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:42 crc kubenswrapper[4685]: E1013 08:45:42.506310 4685 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.524172 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.524215 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.524225 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.524242 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.524258 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:42Z","lastTransitionTime":"2025-10-13T08:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.627196 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.627249 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.627259 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.627277 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.627285 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:42Z","lastTransitionTime":"2025-10-13T08:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.729763 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.729834 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.729853 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.729881 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.729903 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:42Z","lastTransitionTime":"2025-10-13T08:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.786746 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs\") pod \"network-metrics-daemon-nkxgw\" (UID: \"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\") " pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:42 crc kubenswrapper[4685]: E1013 08:45:42.791905 4685 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 08:45:42 crc kubenswrapper[4685]: E1013 08:45:42.792120 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs podName:7cb40c32-d7d4-4a1b-9cda-1de0fab48566 nodeName:}" failed. No retries permitted until 2025-10-13 08:46:14.792073662 +0000 UTC m=+99.939949413 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs") pod "network-metrics-daemon-nkxgw" (UID: "7cb40c32-d7d4-4a1b-9cda-1de0fab48566") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.832145 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.832183 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.832193 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.832209 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.832220 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:42Z","lastTransitionTime":"2025-10-13T08:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.934440 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.934473 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.934484 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.934501 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:42 crc kubenswrapper[4685]: I1013 08:45:42.934511 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:42Z","lastTransitionTime":"2025-10-13T08:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.038242 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.038280 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.038294 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.038311 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.038324 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:43Z","lastTransitionTime":"2025-10-13T08:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.141006 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.141064 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.141078 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.141098 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.141111 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:43Z","lastTransitionTime":"2025-10-13T08:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.244091 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.244183 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.244201 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.244234 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.244251 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:43Z","lastTransitionTime":"2025-10-13T08:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.346852 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.347126 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.347218 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.347349 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.347438 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:43Z","lastTransitionTime":"2025-10-13T08:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.449965 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.450052 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.450066 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.450106 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.450118 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:43Z","lastTransitionTime":"2025-10-13T08:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.501902 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.501994 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:43 crc kubenswrapper[4685]: E1013 08:45:43.502099 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:43 crc kubenswrapper[4685]: E1013 08:45:43.502233 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.554236 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.554292 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.554304 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.554325 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.554337 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:43Z","lastTransitionTime":"2025-10-13T08:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.657485 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.657536 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.657728 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.657745 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.657757 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:43Z","lastTransitionTime":"2025-10-13T08:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.760457 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.760506 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.760517 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.760540 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.760553 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:43Z","lastTransitionTime":"2025-10-13T08:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.863564 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.863604 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.863614 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.863630 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.863639 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:43Z","lastTransitionTime":"2025-10-13T08:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.965625 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.965658 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.965666 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.965680 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:43 crc kubenswrapper[4685]: I1013 08:45:43.965689 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:43Z","lastTransitionTime":"2025-10-13T08:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.067622 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.067652 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.067659 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.067673 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.067682 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:44Z","lastTransitionTime":"2025-10-13T08:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.169823 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.169854 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.169906 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.169941 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.169951 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:44Z","lastTransitionTime":"2025-10-13T08:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.272116 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.272159 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.272169 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.272190 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.272212 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:44Z","lastTransitionTime":"2025-10-13T08:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.375101 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.375164 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.375176 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.375456 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.375479 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:44Z","lastTransitionTime":"2025-10-13T08:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.477781 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.477836 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.477847 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.477862 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.477872 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:44Z","lastTransitionTime":"2025-10-13T08:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.502225 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.502225 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:44 crc kubenswrapper[4685]: E1013 08:45:44.502345 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:44 crc kubenswrapper[4685]: E1013 08:45:44.502401 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.580656 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.580695 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.580708 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.580723 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.580735 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:44Z","lastTransitionTime":"2025-10-13T08:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.683431 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.683467 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.683478 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.683494 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.683505 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:44Z","lastTransitionTime":"2025-10-13T08:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.786702 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.786782 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.786804 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.786841 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.786862 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:44Z","lastTransitionTime":"2025-10-13T08:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.889500 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.889547 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.889560 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.889583 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.889597 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:44Z","lastTransitionTime":"2025-10-13T08:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.971595 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hg77k_ad75c066-910c-49dc-8e8f-94fe04de919d/kube-multus/0.log" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.971694 4685 generic.go:334] "Generic (PLEG): container finished" podID="ad75c066-910c-49dc-8e8f-94fe04de919d" containerID="a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234" exitCode=1 Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.971761 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hg77k" event={"ID":"ad75c066-910c-49dc-8e8f-94fe04de919d","Type":"ContainerDied","Data":"a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234"} Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.972650 4685 scope.go:117] "RemoveContainer" containerID="a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.989460 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:44Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.991906 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.991960 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.991972 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.991991 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:44 crc kubenswrapper[4685]: I1013 08:45:44.992006 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:44Z","lastTransitionTime":"2025-10-13T08:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.018188 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.032698 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.044706 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.075633 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:24Z\\\",\\\"message\\\":\\\".Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1013 08:45:24.446602 6235 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager/kube-controller-manager for network=default are: map[]\\\\nI1013 08:45:24.448730 6235 services_controller.go:443] Built service openshift-kube-controller-manager/kube-controller-manager LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1013 08:45:24.448781 6235 services_controller.go:444] Built service openshift-kube-controller-manager/kube-controller-manager LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1013 08:45:24.448805 6235 services_controller.go:445] Built service openshift-kube-controller-manager/kube-controller-manager LB template configs for network=default: []services.lbConfig(nil)\\\\nF1013 08:45:24.448827 6235 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5kgxk_openshift-ovn-kubernetes(ffa4857b-f741-47d8-b692-74a84be30411)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.094375 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.095515 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.095571 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.095581 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.095601 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.095627 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:45Z","lastTransitionTime":"2025-10-13T08:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.109859 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.124485 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:44Z\\\",\\\"message\\\":\\\"2025-10-13T08:44:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_019f2343-19a3-4d23-8468-38fd31ce9b25\\\\n2025-10-13T08:44:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_019f2343-19a3-4d23-8468-38fd31ce9b25 to /host/opt/cni/bin/\\\\n2025-10-13T08:44:59Z [verbose] multus-daemon started\\\\n2025-10-13T08:44:59Z [verbose] Readiness Indicator file check\\\\n2025-10-13T08:45:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.138070 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.147822 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.159545 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.177214 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.189795 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.197934 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.197955 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.197963 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.197979 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.197989 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:45Z","lastTransitionTime":"2025-10-13T08:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.205836 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.221145 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295380a1-3f1f-4c5e-907c-b0545aa739e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf1b49a1ee1c45ddd3d90831ec5ea3064b05a7ac65ae67ccdfe3dc25ae0e5c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7136baf6f1e826c79c8171ce14480489ff2fab3c1918d6d807535624716b19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p8d8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.234782 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nkxgw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nkxgw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.247955 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d37111fa-463a-422b-916d-c403ce10bf35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d110b9d78594b2f634d07714a305066c53a277b10a9d2ef513e47fc19f52b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://305186b83e0255ae886edebdb2038898352c016ab17ed4a594c7caac64babde0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16b91dfe5b7be866dc336876627a48fb7f5c3d4cfca022c4f7cf2cd448bb9510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.269636 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.301019 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.301055 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.301066 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.301085 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.301097 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:45Z","lastTransitionTime":"2025-10-13T08:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.404031 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.404061 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.404070 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.404084 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.404092 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:45Z","lastTransitionTime":"2025-10-13T08:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.502114 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:45 crc kubenswrapper[4685]: E1013 08:45:45.502278 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.502592 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:45 crc kubenswrapper[4685]: E1013 08:45:45.502682 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.506771 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.506805 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.506814 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.506837 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.506848 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:45Z","lastTransitionTime":"2025-10-13T08:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.517188 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.533963 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.548090 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:44Z\\\",\\\"message\\\":\\\"2025-10-13T08:44:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_019f2343-19a3-4d23-8468-38fd31ce9b25\\\\n2025-10-13T08:44:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_019f2343-19a3-4d23-8468-38fd31ce9b25 to /host/opt/cni/bin/\\\\n2025-10-13T08:44:59Z [verbose] multus-daemon started\\\\n2025-10-13T08:44:59Z [verbose] Readiness Indicator file check\\\\n2025-10-13T08:45:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.568760 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:24Z\\\",\\\"message\\\":\\\".Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1013 08:45:24.446602 6235 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager/kube-controller-manager for network=default are: map[]\\\\nI1013 08:45:24.448730 6235 services_controller.go:443] Built service openshift-kube-controller-manager/kube-controller-manager LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1013 08:45:24.448781 6235 services_controller.go:444] Built service openshift-kube-controller-manager/kube-controller-manager LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1013 08:45:24.448805 6235 services_controller.go:445] Built service openshift-kube-controller-manager/kube-controller-manager LB template configs for network=default: []services.lbConfig(nil)\\\\nF1013 08:45:24.448827 6235 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5kgxk_openshift-ovn-kubernetes(ffa4857b-f741-47d8-b692-74a84be30411)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.589045 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.606828 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.609424 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.609481 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.609500 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.609544 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.609557 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:45Z","lastTransitionTime":"2025-10-13T08:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.620481 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.639625 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.652178 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.666392 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d37111fa-463a-422b-916d-c403ce10bf35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d110b9d78594b2f634d07714a305066c53a277b10a9d2ef513e47fc19f52b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://305186b83e0255ae886edebdb2038898352c016ab17ed4a594c7caac64babde0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16b91dfe5b7be866dc336876627a48fb7f5c3d4cfca022c4f7cf2cd448bb9510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.679812 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.695464 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.708133 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295380a1-3f1f-4c5e-907c-b0545aa739e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf1b49a1ee1c45ddd3d90831ec5ea3064b05a7ac65ae67ccdfe3dc25ae0e5c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7136baf6f1e826c79c8171ce14480489ff2fab3c1918d6d807535624716b19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p8d8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.712320 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.712345 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.712353 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.712366 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.712377 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:45Z","lastTransitionTime":"2025-10-13T08:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.718655 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nkxgw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nkxgw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.736095 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.747330 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.758465 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.768993 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.814615 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.814655 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.814665 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.814682 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.814693 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:45Z","lastTransitionTime":"2025-10-13T08:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.917843 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.918260 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.918271 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.918289 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.918300 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:45Z","lastTransitionTime":"2025-10-13T08:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.977615 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hg77k_ad75c066-910c-49dc-8e8f-94fe04de919d/kube-multus/0.log" Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.977673 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hg77k" event={"ID":"ad75c066-910c-49dc-8e8f-94fe04de919d","Type":"ContainerStarted","Data":"d83eb2c651417915faae46704c55e6fd4132f2dfe079333e33b292e847c15c72"} Oct 13 08:45:45 crc kubenswrapper[4685]: I1013 08:45:45.988240 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:45Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.002450 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:46Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.012506 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:46Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.021874 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.021898 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.021906 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.021945 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.021954 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:46Z","lastTransitionTime":"2025-10-13T08:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.023881 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:46Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.037706 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:46Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.050885 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:46Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.061521 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:46Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.073368 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295380a1-3f1f-4c5e-907c-b0545aa739e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf1b49a1ee1c45ddd3d90831ec5ea3064b05a7ac65ae67ccdfe3dc25ae0e5c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7136baf6f1e826c79c8171ce14480489ff2fab3c1918d6d807535624716b19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p8d8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:46Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.084063 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nkxgw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nkxgw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:46Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.094506 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d37111fa-463a-422b-916d-c403ce10bf35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d110b9d78594b2f634d07714a305066c53a277b10a9d2ef513e47fc19f52b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://305186b83e0255ae886edebdb2038898352c016ab17ed4a594c7caac64babde0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16b91dfe5b7be866dc336876627a48fb7f5c3d4cfca022c4f7cf2cd448bb9510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:46Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.107172 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:46Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.118964 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:46Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.123816 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.123850 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.123861 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.123878 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.123890 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:46Z","lastTransitionTime":"2025-10-13T08:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.140239 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:46Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.153011 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:46Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.166140 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d83eb2c651417915faae46704c55e6fd4132f2dfe079333e33b292e847c15c72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:44Z\\\",\\\"message\\\":\\\"2025-10-13T08:44:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_019f2343-19a3-4d23-8468-38fd31ce9b25\\\\n2025-10-13T08:44:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_019f2343-19a3-4d23-8468-38fd31ce9b25 to /host/opt/cni/bin/\\\\n2025-10-13T08:44:59Z [verbose] multus-daemon started\\\\n2025-10-13T08:44:59Z [verbose] Readiness Indicator file check\\\\n2025-10-13T08:45:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:46Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.187239 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:24Z\\\",\\\"message\\\":\\\".Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1013 08:45:24.446602 6235 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager/kube-controller-manager for network=default are: map[]\\\\nI1013 08:45:24.448730 6235 services_controller.go:443] Built service openshift-kube-controller-manager/kube-controller-manager LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1013 08:45:24.448781 6235 services_controller.go:444] Built service openshift-kube-controller-manager/kube-controller-manager LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1013 08:45:24.448805 6235 services_controller.go:445] Built service openshift-kube-controller-manager/kube-controller-manager LB template configs for network=default: []services.lbConfig(nil)\\\\nF1013 08:45:24.448827 6235 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5kgxk_openshift-ovn-kubernetes(ffa4857b-f741-47d8-b692-74a84be30411)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:46Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.201107 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:46Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.214745 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:46Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.226474 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.226521 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.226534 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.226554 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.226564 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:46Z","lastTransitionTime":"2025-10-13T08:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.329331 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.329378 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.329387 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.329400 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.329410 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:46Z","lastTransitionTime":"2025-10-13T08:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.431678 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.431734 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.431745 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.431765 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.431777 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:46Z","lastTransitionTime":"2025-10-13T08:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.502006 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.502028 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:46 crc kubenswrapper[4685]: E1013 08:45:46.502159 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:46 crc kubenswrapper[4685]: E1013 08:45:46.502325 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.534067 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.534106 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.534116 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.534133 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.534145 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:46Z","lastTransitionTime":"2025-10-13T08:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.637562 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.637598 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.637607 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.637621 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.637630 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:46Z","lastTransitionTime":"2025-10-13T08:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.739887 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.739951 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.739962 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.739977 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.739988 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:46Z","lastTransitionTime":"2025-10-13T08:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.842168 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.842225 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.842236 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.842254 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.842267 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:46Z","lastTransitionTime":"2025-10-13T08:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.944517 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.944557 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.944568 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.944581 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:46 crc kubenswrapper[4685]: I1013 08:45:46.944589 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:46Z","lastTransitionTime":"2025-10-13T08:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.046671 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.046701 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.046709 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.046723 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.046749 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:47Z","lastTransitionTime":"2025-10-13T08:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.149213 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.149242 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.149251 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.149265 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.149274 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:47Z","lastTransitionTime":"2025-10-13T08:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.252025 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.252060 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.252071 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.252088 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.252099 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:47Z","lastTransitionTime":"2025-10-13T08:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.355118 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.355153 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.355161 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.355174 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.355183 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:47Z","lastTransitionTime":"2025-10-13T08:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.457654 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.457693 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.457702 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.457717 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.457726 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:47Z","lastTransitionTime":"2025-10-13T08:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.502043 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.502187 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:47 crc kubenswrapper[4685]: E1013 08:45:47.502265 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:47 crc kubenswrapper[4685]: E1013 08:45:47.502441 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.560682 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.560719 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.560728 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.560743 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.560753 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:47Z","lastTransitionTime":"2025-10-13T08:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.663317 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.663353 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.663362 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.663377 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.663387 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:47Z","lastTransitionTime":"2025-10-13T08:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.766614 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.766664 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.766677 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.766697 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.766708 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:47Z","lastTransitionTime":"2025-10-13T08:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.868476 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.868501 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.868508 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.868521 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.868530 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:47Z","lastTransitionTime":"2025-10-13T08:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.970722 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.970773 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.970786 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.970804 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:47 crc kubenswrapper[4685]: I1013 08:45:47.970817 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:47Z","lastTransitionTime":"2025-10-13T08:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.074808 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.074850 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.074861 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.074877 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.074888 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:48Z","lastTransitionTime":"2025-10-13T08:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.178115 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.178172 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.178185 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.178221 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.178233 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:48Z","lastTransitionTime":"2025-10-13T08:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.284558 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.284599 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.284610 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.284627 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.284637 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:48Z","lastTransitionTime":"2025-10-13T08:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.387107 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.387143 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.387155 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.387170 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.387180 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:48Z","lastTransitionTime":"2025-10-13T08:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.489795 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.489825 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.489834 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.489846 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.489855 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:48Z","lastTransitionTime":"2025-10-13T08:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.501648 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:48 crc kubenswrapper[4685]: E1013 08:45:48.501794 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.502023 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:48 crc kubenswrapper[4685]: E1013 08:45:48.502082 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.512201 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.591777 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.591808 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.591845 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.591859 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.591868 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:48Z","lastTransitionTime":"2025-10-13T08:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.693693 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.693807 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.693864 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.693888 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.693908 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:48Z","lastTransitionTime":"2025-10-13T08:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.796452 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.796491 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.796501 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.796517 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.796526 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:48Z","lastTransitionTime":"2025-10-13T08:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.899362 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.899401 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.899411 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.899427 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:48 crc kubenswrapper[4685]: I1013 08:45:48.899438 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:48Z","lastTransitionTime":"2025-10-13T08:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.001700 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.001729 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.001738 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.001751 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.001760 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:49Z","lastTransitionTime":"2025-10-13T08:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.103766 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.103797 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.103807 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.103826 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.103836 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:49Z","lastTransitionTime":"2025-10-13T08:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.206464 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.206494 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.206502 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.206514 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.206524 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:49Z","lastTransitionTime":"2025-10-13T08:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.308371 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.308409 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.308420 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.308435 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.308443 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:49Z","lastTransitionTime":"2025-10-13T08:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.410883 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.410942 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.410955 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.410973 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.410986 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:49Z","lastTransitionTime":"2025-10-13T08:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.501832 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:49 crc kubenswrapper[4685]: E1013 08:45:49.501998 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.502160 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:49 crc kubenswrapper[4685]: E1013 08:45:49.502205 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.513756 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.513804 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.513813 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.513827 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.513836 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:49Z","lastTransitionTime":"2025-10-13T08:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.616555 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.616589 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.616597 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.616610 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.616620 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:49Z","lastTransitionTime":"2025-10-13T08:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.719257 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.719307 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.719335 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.719350 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.719360 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:49Z","lastTransitionTime":"2025-10-13T08:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.822153 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.822211 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.822221 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.822236 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.822246 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:49Z","lastTransitionTime":"2025-10-13T08:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.924300 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.924382 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.924397 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.924416 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:49 crc kubenswrapper[4685]: I1013 08:45:49.924450 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:49Z","lastTransitionTime":"2025-10-13T08:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.027407 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.027462 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.027473 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.027491 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.027503 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:50Z","lastTransitionTime":"2025-10-13T08:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.131007 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.131042 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.131052 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.131065 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.131074 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:50Z","lastTransitionTime":"2025-10-13T08:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.234051 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.234095 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.234104 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.234119 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.234129 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:50Z","lastTransitionTime":"2025-10-13T08:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.336247 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.336279 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.336288 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.336302 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.336314 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:50Z","lastTransitionTime":"2025-10-13T08:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.438836 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.438868 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.438877 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.438889 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.438899 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:50Z","lastTransitionTime":"2025-10-13T08:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.501807 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.501952 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:50 crc kubenswrapper[4685]: E1013 08:45:50.502057 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:50 crc kubenswrapper[4685]: E1013 08:45:50.502160 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.541617 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.541653 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.541663 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.541677 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.541687 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:50Z","lastTransitionTime":"2025-10-13T08:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.643909 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.643973 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.643986 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.644002 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.644013 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:50Z","lastTransitionTime":"2025-10-13T08:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.746571 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.746602 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.746613 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.746626 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.746636 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:50Z","lastTransitionTime":"2025-10-13T08:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.848554 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.848592 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.848604 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.848623 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.848633 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:50Z","lastTransitionTime":"2025-10-13T08:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.951634 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.951682 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.951693 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.951711 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:50 crc kubenswrapper[4685]: I1013 08:45:50.951723 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:50Z","lastTransitionTime":"2025-10-13T08:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.053877 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.053949 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.053963 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.053984 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.054037 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:51Z","lastTransitionTime":"2025-10-13T08:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.156328 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.156378 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.156391 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.156411 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.156425 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:51Z","lastTransitionTime":"2025-10-13T08:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.258624 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.258675 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.258685 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.258700 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.258712 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:51Z","lastTransitionTime":"2025-10-13T08:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.360829 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.360863 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.360872 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.360890 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.360902 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:51Z","lastTransitionTime":"2025-10-13T08:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.463214 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.463249 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.463260 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.463278 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.463289 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:51Z","lastTransitionTime":"2025-10-13T08:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.501961 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.501960 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:51 crc kubenswrapper[4685]: E1013 08:45:51.502107 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:51 crc kubenswrapper[4685]: E1013 08:45:51.502181 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.565289 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.565325 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.565336 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.565353 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.565364 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:51Z","lastTransitionTime":"2025-10-13T08:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.667517 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.667548 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.667559 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.667575 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.667585 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:51Z","lastTransitionTime":"2025-10-13T08:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.770013 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.770056 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.770064 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.770080 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.770090 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:51Z","lastTransitionTime":"2025-10-13T08:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.872243 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.872276 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.872289 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.872306 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.872317 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:51Z","lastTransitionTime":"2025-10-13T08:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.974834 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.974899 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.974932 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.974948 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:51 crc kubenswrapper[4685]: I1013 08:45:51.974971 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:51Z","lastTransitionTime":"2025-10-13T08:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.077339 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.077384 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.077397 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.077414 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.077424 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:52Z","lastTransitionTime":"2025-10-13T08:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.179329 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.179368 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.179378 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.179393 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.179408 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:52Z","lastTransitionTime":"2025-10-13T08:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.290315 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.290347 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.290356 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.290374 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.290385 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:52Z","lastTransitionTime":"2025-10-13T08:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.391853 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.391880 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.391891 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.391906 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.392167 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:52Z","lastTransitionTime":"2025-10-13T08:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.493892 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.493941 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.493952 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.493968 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.493979 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:52Z","lastTransitionTime":"2025-10-13T08:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.502142 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:52 crc kubenswrapper[4685]: E1013 08:45:52.502227 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.502142 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:52 crc kubenswrapper[4685]: E1013 08:45:52.502315 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.595461 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.595498 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.595509 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.595524 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.595534 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:52Z","lastTransitionTime":"2025-10-13T08:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.697098 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.697124 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.697132 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.697145 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.697156 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:52Z","lastTransitionTime":"2025-10-13T08:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.799927 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.799965 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.799974 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.799990 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.799999 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:52Z","lastTransitionTime":"2025-10-13T08:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.859568 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.859610 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.859621 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.859653 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.859663 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:52Z","lastTransitionTime":"2025-10-13T08:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:52 crc kubenswrapper[4685]: E1013 08:45:52.872745 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:52Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.876750 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.876805 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.876823 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.876844 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.876860 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:52Z","lastTransitionTime":"2025-10-13T08:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:52 crc kubenswrapper[4685]: E1013 08:45:52.893146 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:52Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.896759 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.896803 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.896816 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.896834 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.896847 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:52Z","lastTransitionTime":"2025-10-13T08:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:52 crc kubenswrapper[4685]: E1013 08:45:52.910494 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:52Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.913694 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.913733 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.913743 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.913759 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.913769 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:52Z","lastTransitionTime":"2025-10-13T08:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:52 crc kubenswrapper[4685]: E1013 08:45:52.924412 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:52Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.927425 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.927472 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.927483 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.927497 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.927508 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:52Z","lastTransitionTime":"2025-10-13T08:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:52 crc kubenswrapper[4685]: E1013 08:45:52.938317 4685 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"774b74cd-e138-4e5c-928f-324075caa6f2\\\",\\\"systemUUID\\\":\\\"1ab45c59-0bc0-4269-a6a9-8fdefb164798\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:52Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:52 crc kubenswrapper[4685]: E1013 08:45:52.938437 4685 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.939715 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.939745 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.939755 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.939768 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:52 crc kubenswrapper[4685]: I1013 08:45:52.939778 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:52Z","lastTransitionTime":"2025-10-13T08:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.041973 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.042006 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.042015 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.042028 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.042039 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:53Z","lastTransitionTime":"2025-10-13T08:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.144591 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.144644 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.144654 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.144668 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.144679 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:53Z","lastTransitionTime":"2025-10-13T08:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.247064 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.247112 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.247124 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.247142 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.247158 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:53Z","lastTransitionTime":"2025-10-13T08:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.349505 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.349557 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.349574 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.349597 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.349621 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:53Z","lastTransitionTime":"2025-10-13T08:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.452602 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.452671 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.452698 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.452729 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.452752 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:53Z","lastTransitionTime":"2025-10-13T08:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.501860 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:53 crc kubenswrapper[4685]: E1013 08:45:53.502027 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.502063 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:53 crc kubenswrapper[4685]: E1013 08:45:53.502404 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.502613 4685 scope.go:117] "RemoveContainer" containerID="3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.554788 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.554823 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.554833 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.554850 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.554859 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:53Z","lastTransitionTime":"2025-10-13T08:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.657644 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.657694 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.657704 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.657721 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.657731 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:53Z","lastTransitionTime":"2025-10-13T08:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.760901 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.760963 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.760973 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.760987 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.760998 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:53Z","lastTransitionTime":"2025-10-13T08:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.863781 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.863828 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.863837 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.863855 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.863867 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:53Z","lastTransitionTime":"2025-10-13T08:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.984891 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.984999 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.985012 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.985038 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:53 crc kubenswrapper[4685]: I1013 08:45:53.985050 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:53Z","lastTransitionTime":"2025-10-13T08:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.000238 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5kgxk_ffa4857b-f741-47d8-b692-74a84be30411/ovnkube-controller/2.log" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.002474 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerStarted","Data":"e8877e613e54b392abcce5cb52a14b82c09b70d46953871b3b4be7af5863e89b"} Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.002966 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.014480 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77ca07a4-cb79-405f-a95f-74a57d0e508a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39c698197c289f5a88b55f58360dde69af247a5a1697552450f462c47019be83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8638dfc388b790dc983b512805c3d49c7c25b04ec7708d7d6eb92fd324a6a123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8638dfc388b790dc983b512805c3d49c7c25b04ec7708d7d6eb92fd324a6a123\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:54Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.029074 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d37111fa-463a-422b-916d-c403ce10bf35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d110b9d78594b2f634d07714a305066c53a277b10a9d2ef513e47fc19f52b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://305186b83e0255ae886edebdb2038898352c016ab17ed4a594c7caac64babde0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16b91dfe5b7be866dc336876627a48fb7f5c3d4cfca022c4f7cf2cd448bb9510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:54Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.046865 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:54Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.060230 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:54Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.072471 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295380a1-3f1f-4c5e-907c-b0545aa739e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf1b49a1ee1c45ddd3d90831ec5ea3064b05a7ac65ae67ccdfe3dc25ae0e5c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7136baf6f1e826c79c8171ce14480489ff2fab3c1918d6d807535624716b19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p8d8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:54Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.083405 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nkxgw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nkxgw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:54Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.087221 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.087248 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.087256 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.087270 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.087279 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:54Z","lastTransitionTime":"2025-10-13T08:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.101744 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:54Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.115712 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:54Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.130722 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:54Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.140630 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:54Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.155063 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:54Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.170784 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:54Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.184226 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d83eb2c651417915faae46704c55e6fd4132f2dfe079333e33b292e847c15c72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:44Z\\\",\\\"message\\\":\\\"2025-10-13T08:44:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_019f2343-19a3-4d23-8468-38fd31ce9b25\\\\n2025-10-13T08:44:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_019f2343-19a3-4d23-8468-38fd31ce9b25 to /host/opt/cni/bin/\\\\n2025-10-13T08:44:59Z [verbose] multus-daemon started\\\\n2025-10-13T08:44:59Z [verbose] Readiness Indicator file check\\\\n2025-10-13T08:45:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:54Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.189957 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.190003 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.190015 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.190036 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.190046 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:54Z","lastTransitionTime":"2025-10-13T08:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.211902 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8877e613e54b392abcce5cb52a14b82c09b70d46953871b3b4be7af5863e89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:24Z\\\",\\\"message\\\":\\\".Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1013 08:45:24.446602 6235 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager/kube-controller-manager for network=default are: map[]\\\\nI1013 08:45:24.448730 6235 services_controller.go:443] Built service openshift-kube-controller-manager/kube-controller-manager LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1013 08:45:24.448781 6235 services_controller.go:444] Built service openshift-kube-controller-manager/kube-controller-manager LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1013 08:45:24.448805 6235 services_controller.go:445] Built service openshift-kube-controller-manager/kube-controller-manager LB template configs for network=default: []services.lbConfig(nil)\\\\nF1013 08:45:24.448827 6235 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:54Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.226007 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:54Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.238995 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:54Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.251079 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:54Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.267165 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:54Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.279245 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:54Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.292081 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.292215 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.292228 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.292243 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.292254 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:54Z","lastTransitionTime":"2025-10-13T08:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.394618 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.394661 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.394670 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.394689 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.394699 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:54Z","lastTransitionTime":"2025-10-13T08:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.496857 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.496907 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.496939 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.496965 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.496976 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:54Z","lastTransitionTime":"2025-10-13T08:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.502172 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.502221 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:54 crc kubenswrapper[4685]: E1013 08:45:54.502279 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:54 crc kubenswrapper[4685]: E1013 08:45:54.502328 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.599402 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.599452 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.599465 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.599480 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.599492 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:54Z","lastTransitionTime":"2025-10-13T08:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.702225 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.702269 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.702303 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.702324 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.702338 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:54Z","lastTransitionTime":"2025-10-13T08:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.804730 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.804762 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.804772 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.804785 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.804794 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:54Z","lastTransitionTime":"2025-10-13T08:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.907068 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.907102 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.907111 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.907128 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:54 crc kubenswrapper[4685]: I1013 08:45:54.907137 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:54Z","lastTransitionTime":"2025-10-13T08:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.006539 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5kgxk_ffa4857b-f741-47d8-b692-74a84be30411/ovnkube-controller/3.log" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.007110 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5kgxk_ffa4857b-f741-47d8-b692-74a84be30411/ovnkube-controller/2.log" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.008593 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.008655 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.008666 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.008680 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.008690 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:55Z","lastTransitionTime":"2025-10-13T08:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.009529 4685 generic.go:334] "Generic (PLEG): container finished" podID="ffa4857b-f741-47d8-b692-74a84be30411" containerID="e8877e613e54b392abcce5cb52a14b82c09b70d46953871b3b4be7af5863e89b" exitCode=1 Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.009568 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerDied","Data":"e8877e613e54b392abcce5cb52a14b82c09b70d46953871b3b4be7af5863e89b"} Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.009600 4685 scope.go:117] "RemoveContainer" containerID="3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.010088 4685 scope.go:117] "RemoveContainer" containerID="e8877e613e54b392abcce5cb52a14b82c09b70d46953871b3b4be7af5863e89b" Oct 13 08:45:55 crc kubenswrapper[4685]: E1013 08:45:55.010242 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5kgxk_openshift-ovn-kubernetes(ffa4857b-f741-47d8-b692-74a84be30411)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" podUID="ffa4857b-f741-47d8-b692-74a84be30411" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.021469 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.040678 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.053396 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.066708 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.083313 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8877e613e54b392abcce5cb52a14b82c09b70d46953871b3b4be7af5863e89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:24Z\\\",\\\"message\\\":\\\".Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1013 08:45:24.446602 6235 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager/kube-controller-manager for network=default are: map[]\\\\nI1013 08:45:24.448730 6235 services_controller.go:443] Built service openshift-kube-controller-manager/kube-controller-manager LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1013 08:45:24.448781 6235 services_controller.go:444] Built service openshift-kube-controller-manager/kube-controller-manager LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1013 08:45:24.448805 6235 services_controller.go:445] Built service openshift-kube-controller-manager/kube-controller-manager LB template configs for network=default: []services.lbConfig(nil)\\\\nF1013 08:45:24.448827 6235 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8877e613e54b392abcce5cb52a14b82c09b70d46953871b3b4be7af5863e89b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:54Z\\\",\\\"message\\\":\\\" 6589 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1013 08:45:54.273034 6589 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1013 08:45:54.277472 6589 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1013 08:45:54.277531 6589 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1013 08:45:54.277543 6589 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1013 08:45:54.277547 6589 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1013 08:45:54.277598 6589 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1013 08:45:54.277612 6589 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1013 08:45:54.277623 6589 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1013 08:45:54.277658 6589 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1013 08:45:54.277686 6589 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1013 08:45:54.277712 6589 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1013 08:45:54.277719 6589 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1013 08:45:54.277756 6589 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1013 08:45:54.277765 6589 handler.go:208] Removed *v1.Node event handler 7\\\\nI1013 08:45:54.277786 6589 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.095606 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.107473 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.111220 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.111266 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.111275 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.111291 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.111300 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:55Z","lastTransitionTime":"2025-10-13T08:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.121261 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d83eb2c651417915faae46704c55e6fd4132f2dfe079333e33b292e847c15c72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:44Z\\\",\\\"message\\\":\\\"2025-10-13T08:44:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_019f2343-19a3-4d23-8468-38fd31ce9b25\\\\n2025-10-13T08:44:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_019f2343-19a3-4d23-8468-38fd31ce9b25 to /host/opt/cni/bin/\\\\n2025-10-13T08:44:59Z [verbose] multus-daemon started\\\\n2025-10-13T08:44:59Z [verbose] Readiness Indicator file check\\\\n2025-10-13T08:45:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.135742 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.146063 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.157215 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.167284 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.175816 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.187909 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.200075 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295380a1-3f1f-4c5e-907c-b0545aa739e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf1b49a1ee1c45ddd3d90831ec5ea3064b05a7ac65ae67ccdfe3dc25ae0e5c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7136baf6f1e826c79c8171ce14480489ff2fab3c1918d6d807535624716b19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p8d8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.211393 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nkxgw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nkxgw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.213297 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.213321 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.213330 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.213343 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.213371 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:55Z","lastTransitionTime":"2025-10-13T08:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.223309 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77ca07a4-cb79-405f-a95f-74a57d0e508a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39c698197c289f5a88b55f58360dde69af247a5a1697552450f462c47019be83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8638dfc388b790dc983b512805c3d49c7c25b04ec7708d7d6eb92fd324a6a123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8638dfc388b790dc983b512805c3d49c7c25b04ec7708d7d6eb92fd324a6a123\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.233992 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d37111fa-463a-422b-916d-c403ce10bf35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d110b9d78594b2f634d07714a305066c53a277b10a9d2ef513e47fc19f52b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://305186b83e0255ae886edebdb2038898352c016ab17ed4a594c7caac64babde0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16b91dfe5b7be866dc336876627a48fb7f5c3d4cfca022c4f7cf2cd448bb9510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.245597 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.315889 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.315940 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.315951 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.315967 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.315978 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:55Z","lastTransitionTime":"2025-10-13T08:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.417875 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.417954 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.417969 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.417987 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.418004 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:55Z","lastTransitionTime":"2025-10-13T08:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.502304 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:55 crc kubenswrapper[4685]: E1013 08:45:55.502447 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.502508 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:55 crc kubenswrapper[4685]: E1013 08:45:55.502663 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.513971 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77ca07a4-cb79-405f-a95f-74a57d0e508a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39c698197c289f5a88b55f58360dde69af247a5a1697552450f462c47019be83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8638dfc388b790dc983b512805c3d49c7c25b04ec7708d7d6eb92fd324a6a123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8638dfc388b790dc983b512805c3d49c7c25b04ec7708d7d6eb92fd324a6a123\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.520382 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.520418 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.520427 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.520440 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.520448 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:55Z","lastTransitionTime":"2025-10-13T08:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.527518 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d37111fa-463a-422b-916d-c403ce10bf35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d110b9d78594b2f634d07714a305066c53a277b10a9d2ef513e47fc19f52b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://305186b83e0255ae886edebdb2038898352c016ab17ed4a594c7caac64babde0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16b91dfe5b7be866dc336876627a48fb7f5c3d4cfca022c4f7cf2cd448bb9510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.544187 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.563550 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.580041 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295380a1-3f1f-4c5e-907c-b0545aa739e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf1b49a1ee1c45ddd3d90831ec5ea3064b05a7ac65ae67ccdfe3dc25ae0e5c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7136baf6f1e826c79c8171ce14480489ff2fab3c1918d6d807535624716b19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p8d8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.598343 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nkxgw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nkxgw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.624731 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.624781 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.624803 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.624828 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.624867 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:55Z","lastTransitionTime":"2025-10-13T08:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.627132 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.645148 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.662730 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.676836 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.690888 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.705661 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.719092 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d83eb2c651417915faae46704c55e6fd4132f2dfe079333e33b292e847c15c72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:44Z\\\",\\\"message\\\":\\\"2025-10-13T08:44:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_019f2343-19a3-4d23-8468-38fd31ce9b25\\\\n2025-10-13T08:44:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_019f2343-19a3-4d23-8468-38fd31ce9b25 to /host/opt/cni/bin/\\\\n2025-10-13T08:44:59Z [verbose] multus-daemon started\\\\n2025-10-13T08:44:59Z [verbose] Readiness Indicator file check\\\\n2025-10-13T08:45:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.726801 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.726841 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.726852 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.726869 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.726879 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:55Z","lastTransitionTime":"2025-10-13T08:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.740538 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8877e613e54b392abcce5cb52a14b82c09b70d46953871b3b4be7af5863e89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3acbace6cda441a09a1889d27876be4cebcb3f0e57c534f708f3d9b247421cb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:24Z\\\",\\\"message\\\":\\\".Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1013 08:45:24.446602 6235 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager/kube-controller-manager for network=default are: map[]\\\\nI1013 08:45:24.448730 6235 services_controller.go:443] Built service openshift-kube-controller-manager/kube-controller-manager LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1013 08:45:24.448781 6235 services_controller.go:444] Built service openshift-kube-controller-manager/kube-controller-manager LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1013 08:45:24.448805 6235 services_controller.go:445] Built service openshift-kube-controller-manager/kube-controller-manager LB template configs for network=default: []services.lbConfig(nil)\\\\nF1013 08:45:24.448827 6235 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8877e613e54b392abcce5cb52a14b82c09b70d46953871b3b4be7af5863e89b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:54Z\\\",\\\"message\\\":\\\" 6589 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1013 08:45:54.273034 6589 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1013 08:45:54.277472 6589 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1013 08:45:54.277531 6589 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1013 08:45:54.277543 6589 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1013 08:45:54.277547 6589 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1013 08:45:54.277598 6589 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1013 08:45:54.277612 6589 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1013 08:45:54.277623 6589 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1013 08:45:54.277658 6589 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1013 08:45:54.277686 6589 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1013 08:45:54.277712 6589 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1013 08:45:54.277719 6589 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1013 08:45:54.277756 6589 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1013 08:45:54.277765 6589 handler.go:208] Removed *v1.Node event handler 7\\\\nI1013 08:45:54.277786 6589 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.757134 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.771025 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.781976 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.796397 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.806777 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:55Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.830331 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.830377 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.830388 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.830402 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.830412 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:55Z","lastTransitionTime":"2025-10-13T08:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.932878 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.932961 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.932979 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.933003 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:55 crc kubenswrapper[4685]: I1013 08:45:55.933017 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:55Z","lastTransitionTime":"2025-10-13T08:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.016666 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5kgxk_ffa4857b-f741-47d8-b692-74a84be30411/ovnkube-controller/3.log" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.021861 4685 scope.go:117] "RemoveContainer" containerID="e8877e613e54b392abcce5cb52a14b82c09b70d46953871b3b4be7af5863e89b" Oct 13 08:45:56 crc kubenswrapper[4685]: E1013 08:45:56.022420 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5kgxk_openshift-ovn-kubernetes(ffa4857b-f741-47d8-b692-74a84be30411)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" podUID="ffa4857b-f741-47d8-b692-74a84be30411" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.038363 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.038401 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.038413 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.038434 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.038446 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:56Z","lastTransitionTime":"2025-10-13T08:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.039274 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"313a0145-dae1-4ae0-a841-f5547a965794\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ef3d70dc27691d6705fdbc326e29190e1cc45a9d554f753e2385fce48ddfdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b1b705c9a8741c6250112f4ce91ab7357761d92e629e7a65a6eb4d06aff78c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://034fdcd583f1937d924ad80d5d2cbaa7b38cd1fef50ac0a2dc1f4f234d783936\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724e9d91948cf4c7edb17d95f58ccf0f66430bfa8234f2dc9390440f78202f0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60a12f41cc0aeb23e92b9dcad3a12275a975426d0ba980797b821070bdf87aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feffbca519194e20ffb58f38c2ebf62f6f0389d0a2a9c12a0ce3df2ef299b6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://656956cfc307459cc2b29b0da4f92b1f7b7c497fc745abd1c5ee676df814b055\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:45:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d48dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zmbv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.049287 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-94gwt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6e04a2b-c6c6-4f6c-bf2e-e0902710fc1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0ff7abbbf66fcad11dce2321fc4e191281f1192a782072c7fcd3fca89dc3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwgtv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-94gwt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.066806 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc128c75f1fbc7063d1602b8ea7379209f258a479f48c3868f99bf4fb1bb78fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68072f3d75e6f541e62bf5ec9ae608d1aa7ad9acaf41b6fb3bc36e9562ffede9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.087084 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.099685 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cs95z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4636423-2e13-4c88-90fc-48972c3e2ea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c23a5685012a30438b6eabab1b203d6eb7f86ca65851a76b6bfd03792d5c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rclvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cs95z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.115498 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a0b56093c10d4b712af06b48b17b019f14c3ce1ed28a0e784736d2ffa6d382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.129513 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"295380a1-3f1f-4c5e-907c-b0545aa739e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf1b49a1ee1c45ddd3d90831ec5ea3064b05a7ac65ae67ccdfe3dc25ae0e5c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7136baf6f1e826c79c8171ce14480489ff2fab3c1918d6d807535624716b19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jn5qz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p8d8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.141994 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.142049 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.142067 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.142095 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.142115 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:56Z","lastTransitionTime":"2025-10-13T08:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.142611 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nkxgw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prj95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:45:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nkxgw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.154257 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77ca07a4-cb79-405f-a95f-74a57d0e508a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39c698197c289f5a88b55f58360dde69af247a5a1697552450f462c47019be83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8638dfc388b790dc983b512805c3d49c7c25b04ec7708d7d6eb92fd324a6a123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8638dfc388b790dc983b512805c3d49c7c25b04ec7708d7d6eb92fd324a6a123\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.166386 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d37111fa-463a-422b-916d-c403ce10bf35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d110b9d78594b2f634d07714a305066c53a277b10a9d2ef513e47fc19f52b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://305186b83e0255ae886edebdb2038898352c016ab17ed4a594c7caac64babde0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16b91dfe5b7be866dc336876627a48fb7f5c3d4cfca022c4f7cf2cd448bb9510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b094f7ad43aba6f99417b219f501dcb79763fbe8e7f9af48f98557b0db8e59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.180756 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edd33150d071bf9b02cfb5cc70816bc285696d75888535e50d31aadac5ed44ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.195069 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"505637be-a3fb-4b68-bd17-9f0ed875fb3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1884196f65a2115ff49fb2a607f0da7a0cee71ab272dea5f28f3da7e4d9ce9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-72fpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xrvp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.215566 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f81be30-c065-4a1d-994b-d12fb5b83e7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ead6246435b962c0a1fe344be102f6dbabc6d97c9c2ccaa433c9e51fb0ddf59f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a017e220d5271fad254f0a3032453790ac9ec3d0ea9dbdd84c9e7afde007b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ec674b306f332c7aaf368ec0b69993f5474c614058638dba755e0efeee0b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bca09956fff233674a66383dba926d8339fd1ec61a4eb1b56756bc44d5461ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c5d43d7fc7d0c5590c2f382e979e93f14d1d70d1fa3a771fc5243419e412b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://929e0b09c065eacad8cd1f4845c4b5d9e692e98942f42b8b2fa9c76d414e1488\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b9c21469e1a49e78fd1027d749b58af0498505dc868f4b7f36478ff4eb0d275\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9c9e13b53eee4e05bf5f14d2202e7dd1c0aeff4c7f47391c88b9001add3eab4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.235147 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.245126 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.245195 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.245209 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.245230 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.245242 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:56Z","lastTransitionTime":"2025-10-13T08:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.257440 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.278330 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffa4857b-f741-47d8-b692-74a84be30411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8877e613e54b392abcce5cb52a14b82c09b70d46953871b3b4be7af5863e89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8877e613e54b392abcce5cb52a14b82c09b70d46953871b3b4be7af5863e89b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:54Z\\\",\\\"message\\\":\\\" 6589 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1013 08:45:54.273034 6589 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1013 08:45:54.277472 6589 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1013 08:45:54.277531 6589 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1013 08:45:54.277543 6589 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1013 08:45:54.277547 6589 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1013 08:45:54.277598 6589 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1013 08:45:54.277612 6589 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1013 08:45:54.277623 6589 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1013 08:45:54.277658 6589 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1013 08:45:54.277686 6589 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1013 08:45:54.277712 6589 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1013 08:45:54.277719 6589 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1013 08:45:54.277756 6589 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1013 08:45:54.277765 6589 handler.go:208] Removed *v1.Node event handler 7\\\\nI1013 08:45:54.277786 6589 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:45:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5kgxk_openshift-ovn-kubernetes(ffa4857b-f741-47d8-b692-74a84be30411)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4hsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5kgxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.296286 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa3fb140-9184-4c4b-95b1-1496a7f3c73e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81c63200eb86e82430868a203d7aaa47c37b5b6b7ed59637bcaefe2974cc775e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://220e2efe2eb1bb5ff7f720acb543b0b567c24c14cd7f327c777e75e7212abd9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://174efbd2a30189f6920017f8d5187ab9844e04c32fca37590e6cff1ff38d27d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1f5093a3c985dc7d6f04ed5a6e69bd287e59f42074ed91785afda8bd308cc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16289ada23734792708effbc59c9f8527b078c56b77e2c3ff1478a0e6e23294f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-13T08:44:54Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760345079\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760345079\\\\\\\\\\\\\\\" (2025-10-13 07:44:39 +0000 UTC to 2026-10-13 07:44:39 +0000 UTC (now=2025-10-13 08:44:54.572815332 +0000 UTC))\\\\\\\"\\\\nI1013 08:44:54.572851 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1013 08:44:54.572869 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1013 08:44:54.573058 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573093 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1013 08:44:54.573120 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4234923297/tls.crt::/tmp/serving-cert-4234923297/tls.key\\\\\\\"\\\\nI1013 08:44:54.573308 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1013 08:44:54.573591 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573605 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1013 08:44:54.573624 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1013 08:44:54.573629 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1013 08:44:54.573710 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1013 08:44:54.573716 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1013 08:44:54.574904 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461eb957c2f2ed48e6224c8189a27cd7e80b9a9b9718adf15d993297761206b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://050716fbd96dc4ec566c6a97e907053b68a7ae9350e0fcf077ba25edf9ac30ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T08:44:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.315813 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa48ef40-1409-404e-a65b-5d551383efcc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a902e70927be7ece8bdc07f84ba5ec5d3430569c811b4e804843e86b2c0c367\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://443cddade560ede2d2f97594a6eb213417034398cc5fa84f8488f57f3efbeeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63156e5c3dbcd748bb27e8ffc9068795f92d0934bbb0e127d99006acc2316ae0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28aaacc5a84c4972972ec7516c99e8ba3551c6d99c83157092140fa9b6a5d525\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:44:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.332738 4685 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hg77k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad75c066-910c-49dc-8e8f-94fe04de919d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:44:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T08:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d83eb2c651417915faae46704c55e6fd4132f2dfe079333e33b292e847c15c72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T08:45:44Z\\\",\\\"message\\\":\\\"2025-10-13T08:44:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_019f2343-19a3-4d23-8468-38fd31ce9b25\\\\n2025-10-13T08:44:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_019f2343-19a3-4d23-8468-38fd31ce9b25 to /host/opt/cni/bin/\\\\n2025-10-13T08:44:59Z [verbose] multus-daemon started\\\\n2025-10-13T08:44:59Z [verbose] Readiness Indicator file check\\\\n2025-10-13T08:45:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T08:44:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T08:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdr66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T08:44:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hg77k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T08:45:56Z is after 2025-08-24T17:21:41Z" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.348774 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.348820 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.348833 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.348859 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.348871 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:56Z","lastTransitionTime":"2025-10-13T08:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.451292 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.451353 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.451367 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.451406 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.451419 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:56Z","lastTransitionTime":"2025-10-13T08:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.502172 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.502242 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:56 crc kubenswrapper[4685]: E1013 08:45:56.502360 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:56 crc kubenswrapper[4685]: E1013 08:45:56.502502 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.554420 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.554483 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.554504 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.554532 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.554550 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:56Z","lastTransitionTime":"2025-10-13T08:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.658024 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.658081 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.658095 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.658117 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.658133 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:56Z","lastTransitionTime":"2025-10-13T08:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.760801 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.760855 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.760869 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.760889 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.760903 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:56Z","lastTransitionTime":"2025-10-13T08:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.863944 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.864013 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.864037 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.864067 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.864087 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:56Z","lastTransitionTime":"2025-10-13T08:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.967316 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.967705 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.967893 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.968083 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:56 crc kubenswrapper[4685]: I1013 08:45:56.968286 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:56Z","lastTransitionTime":"2025-10-13T08:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.071252 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.071644 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.071714 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.071809 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.071902 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:57Z","lastTransitionTime":"2025-10-13T08:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.176071 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.176144 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.176161 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.176194 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.176216 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:57Z","lastTransitionTime":"2025-10-13T08:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.283538 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.283581 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.283593 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.283615 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.283627 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:57Z","lastTransitionTime":"2025-10-13T08:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.387418 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.387496 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.387515 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.387582 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.387603 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:57Z","lastTransitionTime":"2025-10-13T08:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.491271 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.491595 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.491671 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.491750 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.491827 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:57Z","lastTransitionTime":"2025-10-13T08:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.502191 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.502215 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:57 crc kubenswrapper[4685]: E1013 08:45:57.502720 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:57 crc kubenswrapper[4685]: E1013 08:45:57.502471 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.594890 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.594959 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.594971 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.594990 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.595001 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:57Z","lastTransitionTime":"2025-10-13T08:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.698835 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.698911 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.698940 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.698959 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.698970 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:57Z","lastTransitionTime":"2025-10-13T08:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.802754 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.802801 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.802810 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.802831 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.802842 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:57Z","lastTransitionTime":"2025-10-13T08:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.942396 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.942784 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.942796 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.942814 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:57 crc kubenswrapper[4685]: I1013 08:45:57.942825 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:57Z","lastTransitionTime":"2025-10-13T08:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.046543 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.046614 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.046637 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.046670 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.046694 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:58Z","lastTransitionTime":"2025-10-13T08:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.149632 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.149685 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.149702 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.149728 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.149745 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:58Z","lastTransitionTime":"2025-10-13T08:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.251941 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.251984 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.251995 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.252013 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.252023 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:58Z","lastTransitionTime":"2025-10-13T08:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.353882 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.354224 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.354291 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.354373 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.354448 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:58Z","lastTransitionTime":"2025-10-13T08:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.456875 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.456965 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.456983 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.457036 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.457057 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:58Z","lastTransitionTime":"2025-10-13T08:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.501703 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:58 crc kubenswrapper[4685]: E1013 08:45:58.501840 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.501837 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:45:58 crc kubenswrapper[4685]: E1013 08:45:58.501957 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.559738 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.559776 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.559786 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.559802 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.559811 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:58Z","lastTransitionTime":"2025-10-13T08:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.662363 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.662412 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.662422 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.662439 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.662450 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:58Z","lastTransitionTime":"2025-10-13T08:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.764847 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.764902 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.764948 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.764971 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.764985 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:58Z","lastTransitionTime":"2025-10-13T08:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.868713 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.868969 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.869050 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.869130 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.869269 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:58Z","lastTransitionTime":"2025-10-13T08:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.971887 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.971956 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.971967 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.971982 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:58 crc kubenswrapper[4685]: I1013 08:45:58.971991 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:58Z","lastTransitionTime":"2025-10-13T08:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.074216 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.074271 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.074282 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.074298 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.074309 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:59Z","lastTransitionTime":"2025-10-13T08:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.176906 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.176953 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.176969 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.176986 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.176997 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:59Z","lastTransitionTime":"2025-10-13T08:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.261205 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.261355 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:59 crc kubenswrapper[4685]: E1013 08:45:59.261420 4685 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 08:45:59 crc kubenswrapper[4685]: E1013 08:45:59.261472 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:03.261418236 +0000 UTC m=+148.409293997 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:45:59 crc kubenswrapper[4685]: E1013 08:45:59.261549 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 08:47:03.261533159 +0000 UTC m=+148.409408930 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.261542 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:59 crc kubenswrapper[4685]: E1013 08:45:59.261632 4685 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 08:45:59 crc kubenswrapper[4685]: E1013 08:45:59.261700 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 08:47:03.261680324 +0000 UTC m=+148.409556275 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.279810 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.279852 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.279863 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.279882 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.279894 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:59Z","lastTransitionTime":"2025-10-13T08:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.363434 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.363504 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:45:59 crc kubenswrapper[4685]: E1013 08:45:59.363625 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 08:45:59 crc kubenswrapper[4685]: E1013 08:45:59.363653 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 08:45:59 crc kubenswrapper[4685]: E1013 08:45:59.363626 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 08:45:59 crc kubenswrapper[4685]: E1013 08:45:59.363683 4685 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 08:45:59 crc kubenswrapper[4685]: E1013 08:45:59.363694 4685 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:45:59 crc kubenswrapper[4685]: E1013 08:45:59.363738 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-13 08:47:03.363723354 +0000 UTC m=+148.511599115 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:45:59 crc kubenswrapper[4685]: E1013 08:45:59.363668 4685 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:45:59 crc kubenswrapper[4685]: E1013 08:45:59.364112 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-13 08:47:03.364045174 +0000 UTC m=+148.511921095 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.382075 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.382109 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.382118 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.382131 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.382140 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:59Z","lastTransitionTime":"2025-10-13T08:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.484952 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.485064 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.485083 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.485112 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.485132 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:59Z","lastTransitionTime":"2025-10-13T08:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.502471 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.502541 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:45:59 crc kubenswrapper[4685]: E1013 08:45:59.502612 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:45:59 crc kubenswrapper[4685]: E1013 08:45:59.503895 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.588065 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.588137 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.588147 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.588164 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.588177 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:59Z","lastTransitionTime":"2025-10-13T08:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.692255 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.692298 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.692311 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.692330 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.692343 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:59Z","lastTransitionTime":"2025-10-13T08:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.795501 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.795555 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.795567 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.795589 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.795606 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:59Z","lastTransitionTime":"2025-10-13T08:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.898714 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.898787 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.898807 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.898836 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:45:59 crc kubenswrapper[4685]: I1013 08:45:59.898864 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:45:59Z","lastTransitionTime":"2025-10-13T08:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.001847 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.001929 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.001941 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.001958 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.001972 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:00Z","lastTransitionTime":"2025-10-13T08:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.105593 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.105633 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.105644 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.105662 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.105672 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:00Z","lastTransitionTime":"2025-10-13T08:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.209551 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.209619 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.209635 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.209659 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.209675 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:00Z","lastTransitionTime":"2025-10-13T08:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.312023 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.312086 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.312100 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.312123 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.312139 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:00Z","lastTransitionTime":"2025-10-13T08:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.416192 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.416258 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.416274 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.416300 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.416318 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:00Z","lastTransitionTime":"2025-10-13T08:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.502706 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:00 crc kubenswrapper[4685]: E1013 08:46:00.502987 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.503047 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:00 crc kubenswrapper[4685]: E1013 08:46:00.503323 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.525650 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.525722 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.525742 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.525769 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.525792 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:00Z","lastTransitionTime":"2025-10-13T08:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.629219 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.629266 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.629278 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.629298 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.629312 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:00Z","lastTransitionTime":"2025-10-13T08:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.732628 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.732772 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.732795 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.732827 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.732848 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:00Z","lastTransitionTime":"2025-10-13T08:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.836280 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.836335 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.836349 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.836373 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.836393 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:00Z","lastTransitionTime":"2025-10-13T08:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.940613 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.940704 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.940730 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.940767 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:00 crc kubenswrapper[4685]: I1013 08:46:00.940794 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:00Z","lastTransitionTime":"2025-10-13T08:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.043620 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.043707 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.043724 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.043753 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.043771 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:01Z","lastTransitionTime":"2025-10-13T08:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.147462 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.147549 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.147572 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.147611 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.147629 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:01Z","lastTransitionTime":"2025-10-13T08:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.250724 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.250786 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.250804 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.250832 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.250848 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:01Z","lastTransitionTime":"2025-10-13T08:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.353568 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.353640 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.353652 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.353670 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.353680 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:01Z","lastTransitionTime":"2025-10-13T08:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.456253 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.456302 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.456311 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.456329 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.456342 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:01Z","lastTransitionTime":"2025-10-13T08:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.502329 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.502329 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:01 crc kubenswrapper[4685]: E1013 08:46:01.502562 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:01 crc kubenswrapper[4685]: E1013 08:46:01.502770 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.559603 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.559653 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.559668 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.559693 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.559712 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:01Z","lastTransitionTime":"2025-10-13T08:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.662439 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.662475 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.662484 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.662498 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.662508 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:01Z","lastTransitionTime":"2025-10-13T08:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.766898 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.766980 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.766995 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.767021 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.767036 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:01Z","lastTransitionTime":"2025-10-13T08:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.870806 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.870895 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.870946 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.870980 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.870999 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:01Z","lastTransitionTime":"2025-10-13T08:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.976992 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.977057 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.977067 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.977083 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:01 crc kubenswrapper[4685]: I1013 08:46:01.977094 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:01Z","lastTransitionTime":"2025-10-13T08:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.080512 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.080574 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.080593 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.080622 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.080642 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:02Z","lastTransitionTime":"2025-10-13T08:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.184500 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.184578 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.184604 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.184638 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.184667 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:02Z","lastTransitionTime":"2025-10-13T08:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.287675 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.287757 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.287784 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.287819 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.287859 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:02Z","lastTransitionTime":"2025-10-13T08:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.391435 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.391515 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.391548 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.391586 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.391613 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:02Z","lastTransitionTime":"2025-10-13T08:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.494469 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.494534 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.494546 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.494566 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.494580 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:02Z","lastTransitionTime":"2025-10-13T08:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.501675 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.501677 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:02 crc kubenswrapper[4685]: E1013 08:46:02.501870 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:02 crc kubenswrapper[4685]: E1013 08:46:02.501994 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.597522 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.597576 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.597591 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.597615 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.597630 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:02Z","lastTransitionTime":"2025-10-13T08:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.700812 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.700903 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.700976 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.701018 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.701044 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:02Z","lastTransitionTime":"2025-10-13T08:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.804375 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.804433 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.804446 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.804508 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.804525 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:02Z","lastTransitionTime":"2025-10-13T08:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.907164 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.907220 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.907232 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.907252 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:02 crc kubenswrapper[4685]: I1013 08:46:02.907266 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:02Z","lastTransitionTime":"2025-10-13T08:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.010027 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.010067 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.010077 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.010094 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.010108 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:03Z","lastTransitionTime":"2025-10-13T08:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.027357 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.027387 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.027401 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.027414 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.027423 4685 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T08:46:03Z","lastTransitionTime":"2025-10-13T08:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.093397 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-g4cwc"] Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.094181 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g4cwc" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.097112 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.097209 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.097416 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.097641 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.132087 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=69.132053994 podStartE2EDuration="1m9.132053994s" podCreationTimestamp="2025-10-13 08:44:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:03.129424176 +0000 UTC m=+88.277299957" watchObservedRunningTime="2025-10-13 08:46:03.132053994 +0000 UTC m=+88.279929765" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.193113 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podStartSLOduration=67.193081238 podStartE2EDuration="1m7.193081238s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:03.175569882 +0000 UTC m=+88.323445683" watchObservedRunningTime="2025-10-13 08:46:03.193081238 +0000 UTC m=+88.340956999" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.209201 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9cce0eb4-2802-4eab-9463-0450302a758f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-g4cwc\" (UID: \"9cce0eb4-2802-4eab-9463-0450302a758f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g4cwc" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.209333 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9cce0eb4-2802-4eab-9463-0450302a758f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-g4cwc\" (UID: \"9cce0eb4-2802-4eab-9463-0450302a758f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g4cwc" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.209366 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9cce0eb4-2802-4eab-9463-0450302a758f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-g4cwc\" (UID: \"9cce0eb4-2802-4eab-9463-0450302a758f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g4cwc" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.209400 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9cce0eb4-2802-4eab-9463-0450302a758f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-g4cwc\" (UID: \"9cce0eb4-2802-4eab-9463-0450302a758f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g4cwc" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.209426 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9cce0eb4-2802-4eab-9463-0450302a758f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-g4cwc\" (UID: \"9cce0eb4-2802-4eab-9463-0450302a758f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g4cwc" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.209527 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=69.20951255 podStartE2EDuration="1m9.20951255s" podCreationTimestamp="2025-10-13 08:44:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:03.208531332 +0000 UTC m=+88.356407113" watchObservedRunningTime="2025-10-13 08:46:03.20951255 +0000 UTC m=+88.357388301" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.229189 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=64.229166728 podStartE2EDuration="1m4.229166728s" podCreationTimestamp="2025-10-13 08:44:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:03.229096946 +0000 UTC m=+88.376972717" watchObservedRunningTime="2025-10-13 08:46:03.229166728 +0000 UTC m=+88.377042489" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.245427 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-hg77k" podStartSLOduration=67.245397885 podStartE2EDuration="1m7.245397885s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:03.245033655 +0000 UTC m=+88.392909426" watchObservedRunningTime="2025-10-13 08:46:03.245397885 +0000 UTC m=+88.393273646" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.283132 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-94gwt" podStartSLOduration=67.283101165 podStartE2EDuration="1m7.283101165s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:03.282429554 +0000 UTC m=+88.430305325" watchObservedRunningTime="2025-10-13 08:46:03.283101165 +0000 UTC m=+88.430976926" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.310053 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9cce0eb4-2802-4eab-9463-0450302a758f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-g4cwc\" (UID: \"9cce0eb4-2802-4eab-9463-0450302a758f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g4cwc" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.310120 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9cce0eb4-2802-4eab-9463-0450302a758f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-g4cwc\" (UID: \"9cce0eb4-2802-4eab-9463-0450302a758f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g4cwc" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.310141 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9cce0eb4-2802-4eab-9463-0450302a758f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-g4cwc\" (UID: \"9cce0eb4-2802-4eab-9463-0450302a758f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g4cwc" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.310166 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9cce0eb4-2802-4eab-9463-0450302a758f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-g4cwc\" (UID: \"9cce0eb4-2802-4eab-9463-0450302a758f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g4cwc" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.310222 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9cce0eb4-2802-4eab-9463-0450302a758f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-g4cwc\" (UID: \"9cce0eb4-2802-4eab-9463-0450302a758f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g4cwc" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.310294 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9cce0eb4-2802-4eab-9463-0450302a758f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-g4cwc\" (UID: \"9cce0eb4-2802-4eab-9463-0450302a758f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g4cwc" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.310581 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9cce0eb4-2802-4eab-9463-0450302a758f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-g4cwc\" (UID: \"9cce0eb4-2802-4eab-9463-0450302a758f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g4cwc" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.311523 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9cce0eb4-2802-4eab-9463-0450302a758f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-g4cwc\" (UID: \"9cce0eb4-2802-4eab-9463-0450302a758f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g4cwc" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.329781 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9cce0eb4-2802-4eab-9463-0450302a758f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-g4cwc\" (UID: \"9cce0eb4-2802-4eab-9463-0450302a758f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g4cwc" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.332135 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-cs95z" podStartSLOduration=67.332124345 podStartE2EDuration="1m7.332124345s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:03.332086774 +0000 UTC m=+88.479962545" watchObservedRunningTime="2025-10-13 08:46:03.332124345 +0000 UTC m=+88.480000106" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.340058 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9cce0eb4-2802-4eab-9463-0450302a758f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-g4cwc\" (UID: \"9cce0eb4-2802-4eab-9463-0450302a758f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g4cwc" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.352685 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-zmbv2" podStartSLOduration=67.35265446 podStartE2EDuration="1m7.35265446s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:03.35236459 +0000 UTC m=+88.500240371" watchObservedRunningTime="2025-10-13 08:46:03.35265446 +0000 UTC m=+88.500530221" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.368427 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p8d8r" podStartSLOduration=67.368390142 podStartE2EDuration="1m7.368390142s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:03.367059993 +0000 UTC m=+88.514935744" watchObservedRunningTime="2025-10-13 08:46:03.368390142 +0000 UTC m=+88.516265903" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.397045 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=15.397025223 podStartE2EDuration="15.397025223s" podCreationTimestamp="2025-10-13 08:45:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:03.396293982 +0000 UTC m=+88.544169743" watchObservedRunningTime="2025-10-13 08:46:03.397025223 +0000 UTC m=+88.544900984" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.411654 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g4cwc" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.447069 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=38.447039914 podStartE2EDuration="38.447039914s" podCreationTimestamp="2025-10-13 08:45:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:03.42069631 +0000 UTC m=+88.568572071" watchObservedRunningTime="2025-10-13 08:46:03.447039914 +0000 UTC m=+88.594915675" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.501854 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:03 crc kubenswrapper[4685]: I1013 08:46:03.501860 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:03 crc kubenswrapper[4685]: E1013 08:46:03.502080 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:03 crc kubenswrapper[4685]: E1013 08:46:03.502165 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:04 crc kubenswrapper[4685]: I1013 08:46:04.051430 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g4cwc" event={"ID":"9cce0eb4-2802-4eab-9463-0450302a758f","Type":"ContainerStarted","Data":"4569c02841cd26fd8f934cc672d4a67ebe95541aa32d94a82b03d9991f5cb2df"} Oct 13 08:46:04 crc kubenswrapper[4685]: I1013 08:46:04.051501 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g4cwc" event={"ID":"9cce0eb4-2802-4eab-9463-0450302a758f","Type":"ContainerStarted","Data":"db76c5f9d5a56fdfeed837ae217fef030f9e2f0a4a776180ce369d4888e03ab0"} Oct 13 08:46:04 crc kubenswrapper[4685]: I1013 08:46:04.502670 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:04 crc kubenswrapper[4685]: I1013 08:46:04.502670 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:04 crc kubenswrapper[4685]: E1013 08:46:04.502856 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:04 crc kubenswrapper[4685]: E1013 08:46:04.502987 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:05 crc kubenswrapper[4685]: I1013 08:46:05.502726 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:05 crc kubenswrapper[4685]: E1013 08:46:05.504355 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:05 crc kubenswrapper[4685]: I1013 08:46:05.504467 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:05 crc kubenswrapper[4685]: E1013 08:46:05.504671 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:06 crc kubenswrapper[4685]: I1013 08:46:06.501872 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:06 crc kubenswrapper[4685]: E1013 08:46:06.502244 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:06 crc kubenswrapper[4685]: I1013 08:46:06.502032 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:06 crc kubenswrapper[4685]: E1013 08:46:06.502449 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:07 crc kubenswrapper[4685]: I1013 08:46:07.502611 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:07 crc kubenswrapper[4685]: E1013 08:46:07.502959 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:07 crc kubenswrapper[4685]: I1013 08:46:07.503624 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:07 crc kubenswrapper[4685]: E1013 08:46:07.503950 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:08 crc kubenswrapper[4685]: I1013 08:46:08.502518 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:08 crc kubenswrapper[4685]: I1013 08:46:08.502572 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:08 crc kubenswrapper[4685]: E1013 08:46:08.503571 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:08 crc kubenswrapper[4685]: E1013 08:46:08.503653 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:08 crc kubenswrapper[4685]: I1013 08:46:08.503965 4685 scope.go:117] "RemoveContainer" containerID="e8877e613e54b392abcce5cb52a14b82c09b70d46953871b3b4be7af5863e89b" Oct 13 08:46:08 crc kubenswrapper[4685]: E1013 08:46:08.504222 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5kgxk_openshift-ovn-kubernetes(ffa4857b-f741-47d8-b692-74a84be30411)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" podUID="ffa4857b-f741-47d8-b692-74a84be30411" Oct 13 08:46:09 crc kubenswrapper[4685]: I1013 08:46:09.501644 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:09 crc kubenswrapper[4685]: I1013 08:46:09.501693 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:09 crc kubenswrapper[4685]: E1013 08:46:09.501753 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:09 crc kubenswrapper[4685]: E1013 08:46:09.501873 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:10 crc kubenswrapper[4685]: I1013 08:46:10.502112 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:10 crc kubenswrapper[4685]: I1013 08:46:10.502117 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:10 crc kubenswrapper[4685]: E1013 08:46:10.502250 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:10 crc kubenswrapper[4685]: E1013 08:46:10.502312 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:11 crc kubenswrapper[4685]: I1013 08:46:11.502191 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:11 crc kubenswrapper[4685]: E1013 08:46:11.502333 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:11 crc kubenswrapper[4685]: I1013 08:46:11.503254 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:11 crc kubenswrapper[4685]: E1013 08:46:11.503524 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:12 crc kubenswrapper[4685]: I1013 08:46:12.502324 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:12 crc kubenswrapper[4685]: E1013 08:46:12.502435 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:12 crc kubenswrapper[4685]: I1013 08:46:12.502623 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:12 crc kubenswrapper[4685]: E1013 08:46:12.502673 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:13 crc kubenswrapper[4685]: I1013 08:46:13.502609 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:13 crc kubenswrapper[4685]: I1013 08:46:13.502874 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:13 crc kubenswrapper[4685]: E1013 08:46:13.502997 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:13 crc kubenswrapper[4685]: E1013 08:46:13.502811 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:14 crc kubenswrapper[4685]: I1013 08:46:14.502327 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:14 crc kubenswrapper[4685]: I1013 08:46:14.502436 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:14 crc kubenswrapper[4685]: E1013 08:46:14.502441 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:14 crc kubenswrapper[4685]: E1013 08:46:14.502678 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:14 crc kubenswrapper[4685]: I1013 08:46:14.850026 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs\") pod \"network-metrics-daemon-nkxgw\" (UID: \"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\") " pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:14 crc kubenswrapper[4685]: E1013 08:46:14.850287 4685 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 08:46:14 crc kubenswrapper[4685]: E1013 08:46:14.850437 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs podName:7cb40c32-d7d4-4a1b-9cda-1de0fab48566 nodeName:}" failed. No retries permitted until 2025-10-13 08:47:18.850401226 +0000 UTC m=+163.998277027 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs") pod "network-metrics-daemon-nkxgw" (UID: "7cb40c32-d7d4-4a1b-9cda-1de0fab48566") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 08:46:15 crc kubenswrapper[4685]: I1013 08:46:15.501768 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:15 crc kubenswrapper[4685]: I1013 08:46:15.501885 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:15 crc kubenswrapper[4685]: E1013 08:46:15.502854 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:15 crc kubenswrapper[4685]: E1013 08:46:15.502948 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:16 crc kubenswrapper[4685]: I1013 08:46:16.502082 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:16 crc kubenswrapper[4685]: E1013 08:46:16.502498 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:16 crc kubenswrapper[4685]: I1013 08:46:16.502082 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:16 crc kubenswrapper[4685]: E1013 08:46:16.502595 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:17 crc kubenswrapper[4685]: I1013 08:46:17.502503 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:17 crc kubenswrapper[4685]: I1013 08:46:17.502565 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:17 crc kubenswrapper[4685]: E1013 08:46:17.502647 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:17 crc kubenswrapper[4685]: E1013 08:46:17.502982 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:18 crc kubenswrapper[4685]: I1013 08:46:18.502585 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:18 crc kubenswrapper[4685]: E1013 08:46:18.502689 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:18 crc kubenswrapper[4685]: I1013 08:46:18.502585 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:18 crc kubenswrapper[4685]: E1013 08:46:18.502748 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:19 crc kubenswrapper[4685]: I1013 08:46:19.502545 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:19 crc kubenswrapper[4685]: I1013 08:46:19.502657 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:19 crc kubenswrapper[4685]: E1013 08:46:19.502687 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:19 crc kubenswrapper[4685]: E1013 08:46:19.502899 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:20 crc kubenswrapper[4685]: I1013 08:46:20.502087 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:20 crc kubenswrapper[4685]: I1013 08:46:20.502132 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:20 crc kubenswrapper[4685]: E1013 08:46:20.502316 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:20 crc kubenswrapper[4685]: E1013 08:46:20.502451 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:21 crc kubenswrapper[4685]: I1013 08:46:21.502593 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:21 crc kubenswrapper[4685]: E1013 08:46:21.502719 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:21 crc kubenswrapper[4685]: I1013 08:46:21.503293 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:21 crc kubenswrapper[4685]: E1013 08:46:21.503508 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:22 crc kubenswrapper[4685]: I1013 08:46:22.501836 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:22 crc kubenswrapper[4685]: E1013 08:46:22.502000 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:22 crc kubenswrapper[4685]: I1013 08:46:22.502012 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:22 crc kubenswrapper[4685]: E1013 08:46:22.502249 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:23 crc kubenswrapper[4685]: I1013 08:46:23.502718 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:23 crc kubenswrapper[4685]: I1013 08:46:23.502774 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:23 crc kubenswrapper[4685]: E1013 08:46:23.503008 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:23 crc kubenswrapper[4685]: E1013 08:46:23.503224 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:23 crc kubenswrapper[4685]: I1013 08:46:23.503876 4685 scope.go:117] "RemoveContainer" containerID="e8877e613e54b392abcce5cb52a14b82c09b70d46953871b3b4be7af5863e89b" Oct 13 08:46:23 crc kubenswrapper[4685]: E1013 08:46:23.504067 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5kgxk_openshift-ovn-kubernetes(ffa4857b-f741-47d8-b692-74a84be30411)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" podUID="ffa4857b-f741-47d8-b692-74a84be30411" Oct 13 08:46:24 crc kubenswrapper[4685]: I1013 08:46:24.502275 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:24 crc kubenswrapper[4685]: E1013 08:46:24.502608 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:24 crc kubenswrapper[4685]: I1013 08:46:24.502304 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:24 crc kubenswrapper[4685]: E1013 08:46:24.502797 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:25 crc kubenswrapper[4685]: I1013 08:46:25.502457 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:25 crc kubenswrapper[4685]: I1013 08:46:25.503077 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:25 crc kubenswrapper[4685]: E1013 08:46:25.503319 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:25 crc kubenswrapper[4685]: E1013 08:46:25.503551 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:26 crc kubenswrapper[4685]: I1013 08:46:26.501768 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:26 crc kubenswrapper[4685]: I1013 08:46:26.501768 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:26 crc kubenswrapper[4685]: E1013 08:46:26.501991 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:26 crc kubenswrapper[4685]: E1013 08:46:26.501889 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:27 crc kubenswrapper[4685]: I1013 08:46:27.503197 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:27 crc kubenswrapper[4685]: I1013 08:46:27.503294 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:27 crc kubenswrapper[4685]: E1013 08:46:27.503442 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:27 crc kubenswrapper[4685]: E1013 08:46:27.503589 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:28 crc kubenswrapper[4685]: I1013 08:46:28.501971 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:28 crc kubenswrapper[4685]: I1013 08:46:28.501997 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:28 crc kubenswrapper[4685]: E1013 08:46:28.502082 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:28 crc kubenswrapper[4685]: E1013 08:46:28.502212 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:29 crc kubenswrapper[4685]: I1013 08:46:29.502251 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:29 crc kubenswrapper[4685]: E1013 08:46:29.502505 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:29 crc kubenswrapper[4685]: I1013 08:46:29.502884 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:29 crc kubenswrapper[4685]: E1013 08:46:29.503069 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:30 crc kubenswrapper[4685]: I1013 08:46:30.502190 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:30 crc kubenswrapper[4685]: I1013 08:46:30.502242 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:30 crc kubenswrapper[4685]: E1013 08:46:30.502367 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:30 crc kubenswrapper[4685]: E1013 08:46:30.502451 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:31 crc kubenswrapper[4685]: I1013 08:46:31.140158 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hg77k_ad75c066-910c-49dc-8e8f-94fe04de919d/kube-multus/1.log" Oct 13 08:46:31 crc kubenswrapper[4685]: I1013 08:46:31.140547 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hg77k_ad75c066-910c-49dc-8e8f-94fe04de919d/kube-multus/0.log" Oct 13 08:46:31 crc kubenswrapper[4685]: I1013 08:46:31.140585 4685 generic.go:334] "Generic (PLEG): container finished" podID="ad75c066-910c-49dc-8e8f-94fe04de919d" containerID="d83eb2c651417915faae46704c55e6fd4132f2dfe079333e33b292e847c15c72" exitCode=1 Oct 13 08:46:31 crc kubenswrapper[4685]: I1013 08:46:31.140613 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hg77k" event={"ID":"ad75c066-910c-49dc-8e8f-94fe04de919d","Type":"ContainerDied","Data":"d83eb2c651417915faae46704c55e6fd4132f2dfe079333e33b292e847c15c72"} Oct 13 08:46:31 crc kubenswrapper[4685]: I1013 08:46:31.140643 4685 scope.go:117] "RemoveContainer" containerID="a194dcc57dc43b3e9971aeaa2547fc98a795260a8df0b51a10b765141aeeb234" Oct 13 08:46:31 crc kubenswrapper[4685]: I1013 08:46:31.140957 4685 scope.go:117] "RemoveContainer" containerID="d83eb2c651417915faae46704c55e6fd4132f2dfe079333e33b292e847c15c72" Oct 13 08:46:31 crc kubenswrapper[4685]: E1013 08:46:31.141095 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-hg77k_openshift-multus(ad75c066-910c-49dc-8e8f-94fe04de919d)\"" pod="openshift-multus/multus-hg77k" podUID="ad75c066-910c-49dc-8e8f-94fe04de919d" Oct 13 08:46:31 crc kubenswrapper[4685]: I1013 08:46:31.164871 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g4cwc" podStartSLOduration=95.164858129 podStartE2EDuration="1m35.164858129s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:04.076601313 +0000 UTC m=+89.224477084" watchObservedRunningTime="2025-10-13 08:46:31.164858129 +0000 UTC m=+116.312733880" Oct 13 08:46:31 crc kubenswrapper[4685]: I1013 08:46:31.501736 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:31 crc kubenswrapper[4685]: E1013 08:46:31.501844 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:31 crc kubenswrapper[4685]: I1013 08:46:31.501706 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:31 crc kubenswrapper[4685]: E1013 08:46:31.502465 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:32 crc kubenswrapper[4685]: I1013 08:46:32.144035 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hg77k_ad75c066-910c-49dc-8e8f-94fe04de919d/kube-multus/1.log" Oct 13 08:46:32 crc kubenswrapper[4685]: I1013 08:46:32.501716 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:32 crc kubenswrapper[4685]: I1013 08:46:32.501768 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:32 crc kubenswrapper[4685]: E1013 08:46:32.501843 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:32 crc kubenswrapper[4685]: E1013 08:46:32.501884 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:33 crc kubenswrapper[4685]: I1013 08:46:33.502335 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:33 crc kubenswrapper[4685]: E1013 08:46:33.502465 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:33 crc kubenswrapper[4685]: I1013 08:46:33.502518 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:33 crc kubenswrapper[4685]: E1013 08:46:33.502846 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:34 crc kubenswrapper[4685]: I1013 08:46:34.501842 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:34 crc kubenswrapper[4685]: I1013 08:46:34.501842 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:34 crc kubenswrapper[4685]: E1013 08:46:34.502002 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:34 crc kubenswrapper[4685]: E1013 08:46:34.502093 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:35 crc kubenswrapper[4685]: E1013 08:46:35.483769 4685 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 13 08:46:35 crc kubenswrapper[4685]: I1013 08:46:35.501986 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:35 crc kubenswrapper[4685]: I1013 08:46:35.502015 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:35 crc kubenswrapper[4685]: E1013 08:46:35.502809 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:35 crc kubenswrapper[4685]: E1013 08:46:35.503036 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:35 crc kubenswrapper[4685]: E1013 08:46:35.602446 4685 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 13 08:46:36 crc kubenswrapper[4685]: I1013 08:46:36.501964 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:36 crc kubenswrapper[4685]: I1013 08:46:36.501968 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:36 crc kubenswrapper[4685]: E1013 08:46:36.502120 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:36 crc kubenswrapper[4685]: E1013 08:46:36.502589 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:36 crc kubenswrapper[4685]: I1013 08:46:36.503040 4685 scope.go:117] "RemoveContainer" containerID="e8877e613e54b392abcce5cb52a14b82c09b70d46953871b3b4be7af5863e89b" Oct 13 08:46:37 crc kubenswrapper[4685]: I1013 08:46:37.165895 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5kgxk_ffa4857b-f741-47d8-b692-74a84be30411/ovnkube-controller/3.log" Oct 13 08:46:37 crc kubenswrapper[4685]: I1013 08:46:37.168803 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerStarted","Data":"379685c6101281fdb70274ed20c9c85e6480bd960816f94d3886064f37a23f5d"} Oct 13 08:46:37 crc kubenswrapper[4685]: I1013 08:46:37.169857 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:46:37 crc kubenswrapper[4685]: I1013 08:46:37.199950 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" podStartSLOduration=101.19990549 podStartE2EDuration="1m41.19990549s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:37.199075416 +0000 UTC m=+122.346951197" watchObservedRunningTime="2025-10-13 08:46:37.19990549 +0000 UTC m=+122.347781251" Oct 13 08:46:37 crc kubenswrapper[4685]: I1013 08:46:37.365405 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-nkxgw"] Oct 13 08:46:37 crc kubenswrapper[4685]: I1013 08:46:37.365743 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:37 crc kubenswrapper[4685]: E1013 08:46:37.365958 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:37 crc kubenswrapper[4685]: I1013 08:46:37.502725 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:37 crc kubenswrapper[4685]: I1013 08:46:37.502801 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:37 crc kubenswrapper[4685]: E1013 08:46:37.502871 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:37 crc kubenswrapper[4685]: E1013 08:46:37.503012 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:38 crc kubenswrapper[4685]: I1013 08:46:38.502637 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:38 crc kubenswrapper[4685]: E1013 08:46:38.502792 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:39 crc kubenswrapper[4685]: I1013 08:46:39.502655 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:39 crc kubenswrapper[4685]: I1013 08:46:39.502709 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:39 crc kubenswrapper[4685]: E1013 08:46:39.502899 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:39 crc kubenswrapper[4685]: I1013 08:46:39.502722 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:39 crc kubenswrapper[4685]: E1013 08:46:39.503010 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:39 crc kubenswrapper[4685]: E1013 08:46:39.503154 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:40 crc kubenswrapper[4685]: I1013 08:46:40.502689 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:40 crc kubenswrapper[4685]: E1013 08:46:40.503022 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:40 crc kubenswrapper[4685]: E1013 08:46:40.604758 4685 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 13 08:46:41 crc kubenswrapper[4685]: I1013 08:46:41.502312 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:41 crc kubenswrapper[4685]: I1013 08:46:41.502391 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:41 crc kubenswrapper[4685]: I1013 08:46:41.502438 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:41 crc kubenswrapper[4685]: E1013 08:46:41.502464 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:41 crc kubenswrapper[4685]: E1013 08:46:41.502602 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:41 crc kubenswrapper[4685]: E1013 08:46:41.502724 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:42 crc kubenswrapper[4685]: I1013 08:46:42.502514 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:42 crc kubenswrapper[4685]: E1013 08:46:42.502887 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:42 crc kubenswrapper[4685]: I1013 08:46:42.503155 4685 scope.go:117] "RemoveContainer" containerID="d83eb2c651417915faae46704c55e6fd4132f2dfe079333e33b292e847c15c72" Oct 13 08:46:43 crc kubenswrapper[4685]: I1013 08:46:43.194300 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hg77k_ad75c066-910c-49dc-8e8f-94fe04de919d/kube-multus/1.log" Oct 13 08:46:43 crc kubenswrapper[4685]: I1013 08:46:43.194368 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hg77k" event={"ID":"ad75c066-910c-49dc-8e8f-94fe04de919d","Type":"ContainerStarted","Data":"566c2c013b9b7e7e48b28168baa279368092660f99ed1276e5c6bb9f48db3032"} Oct 13 08:46:43 crc kubenswrapper[4685]: I1013 08:46:43.502753 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:43 crc kubenswrapper[4685]: I1013 08:46:43.502768 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:43 crc kubenswrapper[4685]: I1013 08:46:43.502906 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:43 crc kubenswrapper[4685]: E1013 08:46:43.503459 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:43 crc kubenswrapper[4685]: E1013 08:46:43.503846 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:43 crc kubenswrapper[4685]: E1013 08:46:43.503616 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:44 crc kubenswrapper[4685]: I1013 08:46:44.502279 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:44 crc kubenswrapper[4685]: E1013 08:46:44.502403 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 08:46:45 crc kubenswrapper[4685]: I1013 08:46:45.502640 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:45 crc kubenswrapper[4685]: I1013 08:46:45.502680 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:45 crc kubenswrapper[4685]: I1013 08:46:45.502808 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:45 crc kubenswrapper[4685]: E1013 08:46:45.504221 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 08:46:45 crc kubenswrapper[4685]: E1013 08:46:45.504581 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 08:46:45 crc kubenswrapper[4685]: E1013 08:46:45.504716 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nkxgw" podUID="7cb40c32-d7d4-4a1b-9cda-1de0fab48566" Oct 13 08:46:46 crc kubenswrapper[4685]: I1013 08:46:46.502364 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:46:46 crc kubenswrapper[4685]: I1013 08:46:46.506989 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 13 08:46:46 crc kubenswrapper[4685]: I1013 08:46:46.507084 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 13 08:46:47 crc kubenswrapper[4685]: I1013 08:46:47.502133 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:46:47 crc kubenswrapper[4685]: I1013 08:46:47.502557 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:46:47 crc kubenswrapper[4685]: I1013 08:46:47.502877 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:46:47 crc kubenswrapper[4685]: I1013 08:46:47.505988 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 13 08:46:47 crc kubenswrapper[4685]: I1013 08:46:47.506331 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 13 08:46:47 crc kubenswrapper[4685]: I1013 08:46:47.509479 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 13 08:46:47 crc kubenswrapper[4685]: I1013 08:46:47.509937 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.518342 4685 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.551466 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kgrk6"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.552946 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.576391 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.576426 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.576514 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.576965 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.577019 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.577653 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.577729 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.578042 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.578047 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.578163 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-54dh8"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.578571 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.578986 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.579701 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.580958 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.581036 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.581110 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.581522 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.582162 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9z47r"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.582476 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.582643 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-lcljr"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.583144 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-lcljr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.585311 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6v84k"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.592483 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-csgb2"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.595968 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6v84k" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.596012 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-csgb2" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.597646 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8ks9s"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.616890 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-8ks9s" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.620062 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-2kbsj"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.620679 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j6kt6"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.621155 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-gx6qj"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.621625 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.622380 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-2kbsj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.622756 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-j6kt6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.623948 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8jzsj"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.624336 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-l74cd"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.624604 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-pbpvq"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.625109 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pbpvq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.625478 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.625788 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.626375 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84w94"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.626845 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pk75f"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.627189 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pk75f" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.627381 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84w94" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.627794 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-9cnpf"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.628432 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-9cnpf" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.635458 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-4lflr"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.636367 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gpr4j"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.636801 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gpr4j" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.636945 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8x7bf"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.637255 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4lflr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.637566 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8x7bf" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.637764 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-r96rq"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.638290 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-r96rq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.640206 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rx99b"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.640672 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-rx99b" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.640683 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qm75h"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.641632 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qm75h" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719156 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a81f8b4-7529-45f7-8ea5-df3252d70653-config\") pod \"controller-manager-879f6c89f-9z47r\" (UID: \"7a81f8b4-7529-45f7-8ea5-df3252d70653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719206 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spclf\" (UniqueName: \"kubernetes.io/projected/e8d90978-a395-4313-bcb2-ca35a303b518-kube-api-access-spclf\") pod \"cluster-samples-operator-665b6dd947-csgb2\" (UID: \"e8d90978-a395-4313-bcb2-ca35a303b518\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-csgb2" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719226 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/887c5808-d4ac-4517-a1b9-1601f79aafa8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-gpr4j\" (UID: \"887c5808-d4ac-4517-a1b9-1601f79aafa8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gpr4j" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719251 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719272 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2f79e15-fea6-4391-a711-53ca225b1857-trusted-ca\") pod \"ingress-operator-5b745b69d9-pbpvq\" (UID: \"c2f79e15-fea6-4391-a711-53ca225b1857\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pbpvq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719310 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2158024-4bf6-457d-809d-c389b8ff2309-serving-cert\") pod \"route-controller-manager-6576b87f9c-ssk8b\" (UID: \"b2158024-4bf6-457d-809d-c389b8ff2309\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719328 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-image-import-ca\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719372 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/76cb8882-50e3-4443-b73e-27dbb83e4173-proxy-tls\") pod \"machine-config-operator-74547568cd-qm75h\" (UID: \"76cb8882-50e3-4443-b73e-27dbb83e4173\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qm75h" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719390 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-serving-cert\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719407 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g7gm\" (UniqueName: \"kubernetes.io/projected/44a7f769-9a31-4555-be16-51e310ac34e3-kube-api-access-8g7gm\") pod \"downloads-7954f5f757-2kbsj\" (UID: \"44a7f769-9a31-4555-be16-51e310ac34e3\") " pod="openshift-console/downloads-7954f5f757-2kbsj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719422 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c96407ea-0763-479d-8020-ace1607bd427-config\") pod \"authentication-operator-69f744f599-lcljr\" (UID: \"c96407ea-0763-479d-8020-ace1607bd427\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lcljr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719437 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bcca1906-64bf-44dc-9307-b8d44d2b506f-metrics-tls\") pod \"dns-operator-744455d44c-9cnpf\" (UID: \"bcca1906-64bf-44dc-9307-b8d44d2b506f\") " pod="openshift-dns-operator/dns-operator-744455d44c-9cnpf" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719455 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-audit-policies\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719469 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719488 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2158024-4bf6-457d-809d-c389b8ff2309-config\") pod \"route-controller-manager-6576b87f9c-ssk8b\" (UID: \"b2158024-4bf6-457d-809d-c389b8ff2309\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719517 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-audit-policies\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719534 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c96407ea-0763-479d-8020-ace1607bd427-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-lcljr\" (UID: \"c96407ea-0763-479d-8020-ace1607bd427\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lcljr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719551 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-config\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719564 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c2f79e15-fea6-4391-a711-53ca225b1857-metrics-tls\") pod \"ingress-operator-5b745b69d9-pbpvq\" (UID: \"c2f79e15-fea6-4391-a711-53ca225b1857\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pbpvq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719579 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b57dcc9c-6002-4c47-ae85-c287659cb679-etcd-client\") pod \"etcd-operator-b45778765-l74cd\" (UID: \"b57dcc9c-6002-4c47-ae85-c287659cb679\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719593 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a185671e-0a3a-4a9e-a884-65b448b1e922-images\") pod \"machine-api-operator-5694c8668f-j6kt6\" (UID: \"a185671e-0a3a-4a9e-a884-65b448b1e922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6kt6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719611 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9whn\" (UniqueName: \"kubernetes.io/projected/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-kube-api-access-x9whn\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719627 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a81f8b4-7529-45f7-8ea5-df3252d70653-serving-cert\") pod \"controller-manager-879f6c89f-9z47r\" (UID: \"7a81f8b4-7529-45f7-8ea5-df3252d70653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719645 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/16484304-6976-45e8-a495-f1d2ee367f0f-console-serving-cert\") pod \"console-f9d7485db-gx6qj\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719673 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25a48963-29f5-4d84-91cf-b80e380ed9f1-config\") pod \"kube-controller-manager-operator-78b949d7b-pk75f\" (UID: \"25a48963-29f5-4d84-91cf-b80e380ed9f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pk75f" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719693 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxr6m\" (UniqueName: \"kubernetes.io/projected/a185671e-0a3a-4a9e-a884-65b448b1e922-kube-api-access-jxr6m\") pod \"machine-api-operator-5694c8668f-j6kt6\" (UID: \"a185671e-0a3a-4a9e-a884-65b448b1e922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6kt6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719717 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/83d9c4d8-10b5-4d1e-a71e-e6e92efd8379-trusted-ca\") pod \"console-operator-58897d9998-8ks9s\" (UID: \"83d9c4d8-10b5-4d1e-a71e-e6e92efd8379\") " pod="openshift-console-operator/console-operator-58897d9998-8ks9s" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719738 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719756 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719774 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-serving-cert\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719791 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719815 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719831 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83d9c4d8-10b5-4d1e-a71e-e6e92efd8379-config\") pod \"console-operator-58897d9998-8ks9s\" (UID: \"83d9c4d8-10b5-4d1e-a71e-e6e92efd8379\") " pod="openshift-console-operator/console-operator-58897d9998-8ks9s" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719847 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gztm5\" (UniqueName: \"kubernetes.io/projected/b9f835b0-16cf-4f05-a8a2-2767a929ca25-kube-api-access-gztm5\") pod \"multus-admission-controller-857f4d67dd-rx99b\" (UID: \"b9f835b0-16cf-4f05-a8a2-2767a929ca25\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rx99b" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719863 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0fe547bf-ea91-4546-a886-c613abeae02a-service-ca-bundle\") pod \"router-default-5444994796-r96rq\" (UID: \"0fe547bf-ea91-4546-a886-c613abeae02a\") " pod="openshift-ingress/router-default-5444994796-r96rq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719878 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7826m\" (UniqueName: \"kubernetes.io/projected/bcca1906-64bf-44dc-9307-b8d44d2b506f-kube-api-access-7826m\") pod \"dns-operator-744455d44c-9cnpf\" (UID: \"bcca1906-64bf-44dc-9307-b8d44d2b506f\") " pod="openshift-dns-operator/dns-operator-744455d44c-9cnpf" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719895 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-audit-dir\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719927 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b9f835b0-16cf-4f05-a8a2-2767a929ca25-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rx99b\" (UID: \"b9f835b0-16cf-4f05-a8a2-2767a929ca25\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rx99b" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719943 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-service-ca\") pod \"console-f9d7485db-gx6qj\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719957 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-trusted-ca-bundle\") pod \"console-f9d7485db-gx6qj\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719972 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-oauth-serving-cert\") pod \"console-f9d7485db-gx6qj\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.719987 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/25a48963-29f5-4d84-91cf-b80e380ed9f1-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pk75f\" (UID: \"25a48963-29f5-4d84-91cf-b80e380ed9f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pk75f" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720009 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a185671e-0a3a-4a9e-a884-65b448b1e922-config\") pod \"machine-api-operator-5694c8668f-j6kt6\" (UID: \"a185671e-0a3a-4a9e-a884-65b448b1e922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6kt6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720027 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720044 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b57dcc9c-6002-4c47-ae85-c287659cb679-config\") pod \"etcd-operator-b45778765-l74cd\" (UID: \"b57dcc9c-6002-4c47-ae85-c287659cb679\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720059 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/16484304-6976-45e8-a495-f1d2ee367f0f-console-oauth-config\") pod \"console-f9d7485db-gx6qj\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720073 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c96407ea-0763-479d-8020-ace1607bd427-service-ca-bundle\") pod \"authentication-operator-69f744f599-lcljr\" (UID: \"c96407ea-0763-479d-8020-ace1607bd427\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lcljr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720089 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xzss\" (UniqueName: \"kubernetes.io/projected/3937dccc-95e5-4aa3-be81-6d12c31f7cb9-kube-api-access-2xzss\") pod \"openshift-controller-manager-operator-756b6f6bc6-6v84k\" (UID: \"3937dccc-95e5-4aa3-be81-6d12c31f7cb9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6v84k" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720107 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/887c5808-d4ac-4517-a1b9-1601f79aafa8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-gpr4j\" (UID: \"887c5808-d4ac-4517-a1b9-1601f79aafa8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gpr4j" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720124 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/76cb8882-50e3-4443-b73e-27dbb83e4173-images\") pod \"machine-config-operator-74547568cd-qm75h\" (UID: \"76cb8882-50e3-4443-b73e-27dbb83e4173\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qm75h" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720139 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-etcd-client\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720152 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7a81f8b4-7529-45f7-8ea5-df3252d70653-client-ca\") pod \"controller-manager-879f6c89f-9z47r\" (UID: \"7a81f8b4-7529-45f7-8ea5-df3252d70653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720167 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jcb9\" (UniqueName: \"kubernetes.io/projected/0fe547bf-ea91-4546-a886-c613abeae02a-kube-api-access-4jcb9\") pod \"router-default-5444994796-r96rq\" (UID: \"0fe547bf-ea91-4546-a886-c613abeae02a\") " pod="openshift-ingress/router-default-5444994796-r96rq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720185 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720201 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpqnp\" (UniqueName: \"kubernetes.io/projected/b2158024-4bf6-457d-809d-c389b8ff2309-kube-api-access-mpqnp\") pod \"route-controller-manager-6576b87f9c-ssk8b\" (UID: \"b2158024-4bf6-457d-809d-c389b8ff2309\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720216 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e8d90978-a395-4313-bcb2-ca35a303b518-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-csgb2\" (UID: \"e8d90978-a395-4313-bcb2-ca35a303b518\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-csgb2" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720231 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5072c69e-c292-478d-9998-8b5237dfc33f-audit-dir\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720247 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720266 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v5tw\" (UniqueName: \"kubernetes.io/projected/c96407ea-0763-479d-8020-ace1607bd427-kube-api-access-5v5tw\") pod \"authentication-operator-69f744f599-lcljr\" (UID: \"c96407ea-0763-479d-8020-ace1607bd427\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lcljr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720288 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720308 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-console-config\") pod \"console-f9d7485db-gx6qj\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720329 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-audit-dir\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720345 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbp54\" (UniqueName: \"kubernetes.io/projected/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-kube-api-access-nbp54\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720365 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z579t\" (UniqueName: \"kubernetes.io/projected/76cb8882-50e3-4443-b73e-27dbb83e4173-kube-api-access-z579t\") pod \"machine-config-operator-74547568cd-qm75h\" (UID: \"76cb8882-50e3-4443-b73e-27dbb83e4173\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qm75h" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720384 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/b57dcc9c-6002-4c47-ae85-c287659cb679-etcd-service-ca\") pod \"etcd-operator-b45778765-l74cd\" (UID: \"b57dcc9c-6002-4c47-ae85-c287659cb679\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720404 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl6bm\" (UniqueName: \"kubernetes.io/projected/b57dcc9c-6002-4c47-ae85-c287659cb679-kube-api-access-pl6bm\") pod \"etcd-operator-b45778765-l74cd\" (UID: \"b57dcc9c-6002-4c47-ae85-c287659cb679\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720424 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k2b5\" (UniqueName: \"kubernetes.io/projected/5072c69e-c292-478d-9998-8b5237dfc33f-kube-api-access-5k2b5\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720443 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-etcd-serving-ca\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720462 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/b57dcc9c-6002-4c47-ae85-c287659cb679-etcd-ca\") pod \"etcd-operator-b45778765-l74cd\" (UID: \"b57dcc9c-6002-4c47-ae85-c287659cb679\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720492 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720511 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c96407ea-0763-479d-8020-ace1607bd427-serving-cert\") pod \"authentication-operator-69f744f599-lcljr\" (UID: \"c96407ea-0763-479d-8020-ace1607bd427\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lcljr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720529 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b57dcc9c-6002-4c47-ae85-c287659cb679-serving-cert\") pod \"etcd-operator-b45778765-l74cd\" (UID: \"b57dcc9c-6002-4c47-ae85-c287659cb679\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720545 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0fe547bf-ea91-4546-a886-c613abeae02a-default-certificate\") pod \"router-default-5444994796-r96rq\" (UID: \"0fe547bf-ea91-4546-a886-c613abeae02a\") " pod="openshift-ingress/router-default-5444994796-r96rq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720567 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh6nw\" (UniqueName: \"kubernetes.io/projected/7a81f8b4-7529-45f7-8ea5-df3252d70653-kube-api-access-nh6nw\") pod \"controller-manager-879f6c89f-9z47r\" (UID: \"7a81f8b4-7529-45f7-8ea5-df3252d70653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720582 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a185671e-0a3a-4a9e-a884-65b448b1e922-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j6kt6\" (UID: \"a185671e-0a3a-4a9e-a884-65b448b1e922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6kt6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720599 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7a81f8b4-7529-45f7-8ea5-df3252d70653-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9z47r\" (UID: \"7a81f8b4-7529-45f7-8ea5-df3252d70653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720620 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3937dccc-95e5-4aa3-be81-6d12c31f7cb9-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-6v84k\" (UID: \"3937dccc-95e5-4aa3-be81-6d12c31f7cb9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6v84k" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720634 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-etcd-client\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720652 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720675 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720694 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-node-pullsecrets\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720713 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0fe547bf-ea91-4546-a886-c613abeae02a-metrics-certs\") pod \"router-default-5444994796-r96rq\" (UID: \"0fe547bf-ea91-4546-a886-c613abeae02a\") " pod="openshift-ingress/router-default-5444994796-r96rq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720733 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgvfj\" (UniqueName: \"kubernetes.io/projected/887c5808-d4ac-4517-a1b9-1601f79aafa8-kube-api-access-jgvfj\") pod \"openshift-apiserver-operator-796bbdcf4f-gpr4j\" (UID: \"887c5808-d4ac-4517-a1b9-1601f79aafa8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gpr4j" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720765 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3937dccc-95e5-4aa3-be81-6d12c31f7cb9-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-6v84k\" (UID: \"3937dccc-95e5-4aa3-be81-6d12c31f7cb9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6v84k" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720787 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/76cb8882-50e3-4443-b73e-27dbb83e4173-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qm75h\" (UID: \"76cb8882-50e3-4443-b73e-27dbb83e4173\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qm75h" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720808 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfltc\" (UniqueName: \"kubernetes.io/projected/c2f79e15-fea6-4391-a711-53ca225b1857-kube-api-access-vfltc\") pod \"ingress-operator-5b745b69d9-pbpvq\" (UID: \"c2f79e15-fea6-4391-a711-53ca225b1857\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pbpvq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720835 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s4dp\" (UniqueName: \"kubernetes.io/projected/83d9c4d8-10b5-4d1e-a71e-e6e92efd8379-kube-api-access-9s4dp\") pod \"console-operator-58897d9998-8ks9s\" (UID: \"83d9c4d8-10b5-4d1e-a71e-e6e92efd8379\") " pod="openshift-console-operator/console-operator-58897d9998-8ks9s" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720855 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-encryption-config\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720874 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0fe547bf-ea91-4546-a886-c613abeae02a-stats-auth\") pod \"router-default-5444994796-r96rq\" (UID: \"0fe547bf-ea91-4546-a886-c613abeae02a\") " pod="openshift-ingress/router-default-5444994796-r96rq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720893 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b2158024-4bf6-457d-809d-c389b8ff2309-client-ca\") pod \"route-controller-manager-6576b87f9c-ssk8b\" (UID: \"b2158024-4bf6-457d-809d-c389b8ff2309\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720927 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-audit\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720947 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c2f79e15-fea6-4391-a711-53ca225b1857-bound-sa-token\") pod \"ingress-operator-5b745b69d9-pbpvq\" (UID: \"c2f79e15-fea6-4391-a711-53ca225b1857\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pbpvq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720967 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hfd8\" (UniqueName: \"kubernetes.io/projected/16484304-6976-45e8-a495-f1d2ee367f0f-kube-api-access-6hfd8\") pod \"console-f9d7485db-gx6qj\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.720986 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83d9c4d8-10b5-4d1e-a71e-e6e92efd8379-serving-cert\") pod \"console-operator-58897d9998-8ks9s\" (UID: \"83d9c4d8-10b5-4d1e-a71e-e6e92efd8379\") " pod="openshift-console-operator/console-operator-58897d9998-8ks9s" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.721008 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.721027 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-encryption-config\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.721047 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25a48963-29f5-4d84-91cf-b80e380ed9f1-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pk75f\" (UID: \"25a48963-29f5-4d84-91cf-b80e380ed9f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pk75f" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.735277 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.735677 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.735856 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.736699 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.736896 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.737090 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.737268 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.738391 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.741583 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-9c4d9"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.747476 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.747727 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.747849 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.748008 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.748174 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.748491 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.748638 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.748788 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.748984 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.749194 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.749791 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.750058 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.750220 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.750363 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.750367 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.750532 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.750681 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.750704 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.750926 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.759504 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c4d9" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.774352 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.775286 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.777050 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.778303 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.779228 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.779427 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.779877 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.796373 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.796432 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.796852 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.797126 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.801721 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.804107 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.804659 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.804825 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.804993 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.805140 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.805288 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.805422 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.805728 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.805872 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.806203 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.806378 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.806534 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.806677 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.806817 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.806977 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.807254 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.807665 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.807847 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.808144 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.810298 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.817284 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.818164 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.818302 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.837153 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.837438 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.838114 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.838331 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.838718 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.841731 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.842258 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.842590 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.842722 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.842856 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.842985 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.843710 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.843905 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.844360 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.844579 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.844767 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.847813 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.848397 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.848523 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ljhpv"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.849074 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.849167 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ljhpv" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.849716 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.849086 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-n8hmz"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.850351 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.850413 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n8hmz" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.851403 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.851601 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.851995 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.852151 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.852285 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.852634 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-audit-dir\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.852662 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbp54\" (UniqueName: \"kubernetes.io/projected/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-kube-api-access-nbp54\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.852681 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z579t\" (UniqueName: \"kubernetes.io/projected/76cb8882-50e3-4443-b73e-27dbb83e4173-kube-api-access-z579t\") pod \"machine-config-operator-74547568cd-qm75h\" (UID: \"76cb8882-50e3-4443-b73e-27dbb83e4173\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qm75h" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.852703 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/b57dcc9c-6002-4c47-ae85-c287659cb679-etcd-service-ca\") pod \"etcd-operator-b45778765-l74cd\" (UID: \"b57dcc9c-6002-4c47-ae85-c287659cb679\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.852721 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl6bm\" (UniqueName: \"kubernetes.io/projected/b57dcc9c-6002-4c47-ae85-c287659cb679-kube-api-access-pl6bm\") pod \"etcd-operator-b45778765-l74cd\" (UID: \"b57dcc9c-6002-4c47-ae85-c287659cb679\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.852740 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k2b5\" (UniqueName: \"kubernetes.io/projected/5072c69e-c292-478d-9998-8b5237dfc33f-kube-api-access-5k2b5\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.852757 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-etcd-serving-ca\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.852775 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/b57dcc9c-6002-4c47-ae85-c287659cb679-etcd-ca\") pod \"etcd-operator-b45778765-l74cd\" (UID: \"b57dcc9c-6002-4c47-ae85-c287659cb679\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.852801 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.852819 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c96407ea-0763-479d-8020-ace1607bd427-serving-cert\") pod \"authentication-operator-69f744f599-lcljr\" (UID: \"c96407ea-0763-479d-8020-ace1607bd427\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lcljr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.852835 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b57dcc9c-6002-4c47-ae85-c287659cb679-serving-cert\") pod \"etcd-operator-b45778765-l74cd\" (UID: \"b57dcc9c-6002-4c47-ae85-c287659cb679\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.852850 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0fe547bf-ea91-4546-a886-c613abeae02a-default-certificate\") pod \"router-default-5444994796-r96rq\" (UID: \"0fe547bf-ea91-4546-a886-c613abeae02a\") " pod="openshift-ingress/router-default-5444994796-r96rq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.852875 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh6nw\" (UniqueName: \"kubernetes.io/projected/7a81f8b4-7529-45f7-8ea5-df3252d70653-kube-api-access-nh6nw\") pod \"controller-manager-879f6c89f-9z47r\" (UID: \"7a81f8b4-7529-45f7-8ea5-df3252d70653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.852892 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a185671e-0a3a-4a9e-a884-65b448b1e922-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j6kt6\" (UID: \"a185671e-0a3a-4a9e-a884-65b448b1e922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6kt6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.852926 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7a81f8b4-7529-45f7-8ea5-df3252d70653-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9z47r\" (UID: \"7a81f8b4-7529-45f7-8ea5-df3252d70653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.852945 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3937dccc-95e5-4aa3-be81-6d12c31f7cb9-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-6v84k\" (UID: \"3937dccc-95e5-4aa3-be81-6d12c31f7cb9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6v84k" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.852960 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-etcd-client\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.852977 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.852997 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853012 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-node-pullsecrets\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853031 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgvfj\" (UniqueName: \"kubernetes.io/projected/887c5808-d4ac-4517-a1b9-1601f79aafa8-kube-api-access-jgvfj\") pod \"openshift-apiserver-operator-796bbdcf4f-gpr4j\" (UID: \"887c5808-d4ac-4517-a1b9-1601f79aafa8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gpr4j" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853049 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0fe547bf-ea91-4546-a886-c613abeae02a-metrics-certs\") pod \"router-default-5444994796-r96rq\" (UID: \"0fe547bf-ea91-4546-a886-c613abeae02a\") " pod="openshift-ingress/router-default-5444994796-r96rq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853067 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3937dccc-95e5-4aa3-be81-6d12c31f7cb9-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-6v84k\" (UID: \"3937dccc-95e5-4aa3-be81-6d12c31f7cb9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6v84k" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853083 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/76cb8882-50e3-4443-b73e-27dbb83e4173-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qm75h\" (UID: \"76cb8882-50e3-4443-b73e-27dbb83e4173\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qm75h" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853102 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfltc\" (UniqueName: \"kubernetes.io/projected/c2f79e15-fea6-4391-a711-53ca225b1857-kube-api-access-vfltc\") pod \"ingress-operator-5b745b69d9-pbpvq\" (UID: \"c2f79e15-fea6-4391-a711-53ca225b1857\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pbpvq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853127 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s4dp\" (UniqueName: \"kubernetes.io/projected/83d9c4d8-10b5-4d1e-a71e-e6e92efd8379-kube-api-access-9s4dp\") pod \"console-operator-58897d9998-8ks9s\" (UID: \"83d9c4d8-10b5-4d1e-a71e-e6e92efd8379\") " pod="openshift-console-operator/console-operator-58897d9998-8ks9s" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853146 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-encryption-config\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853165 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0fe547bf-ea91-4546-a886-c613abeae02a-stats-auth\") pod \"router-default-5444994796-r96rq\" (UID: \"0fe547bf-ea91-4546-a886-c613abeae02a\") " pod="openshift-ingress/router-default-5444994796-r96rq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853185 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83d9c4d8-10b5-4d1e-a71e-e6e92efd8379-serving-cert\") pod \"console-operator-58897d9998-8ks9s\" (UID: \"83d9c4d8-10b5-4d1e-a71e-e6e92efd8379\") " pod="openshift-console-operator/console-operator-58897d9998-8ks9s" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853203 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b2158024-4bf6-457d-809d-c389b8ff2309-client-ca\") pod \"route-controller-manager-6576b87f9c-ssk8b\" (UID: \"b2158024-4bf6-457d-809d-c389b8ff2309\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853223 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-audit\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853242 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c2f79e15-fea6-4391-a711-53ca225b1857-bound-sa-token\") pod \"ingress-operator-5b745b69d9-pbpvq\" (UID: \"c2f79e15-fea6-4391-a711-53ca225b1857\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pbpvq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853262 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hfd8\" (UniqueName: \"kubernetes.io/projected/16484304-6976-45e8-a495-f1d2ee367f0f-kube-api-access-6hfd8\") pod \"console-f9d7485db-gx6qj\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853281 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853299 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-encryption-config\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853318 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25a48963-29f5-4d84-91cf-b80e380ed9f1-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pk75f\" (UID: \"25a48963-29f5-4d84-91cf-b80e380ed9f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pk75f" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853336 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a81f8b4-7529-45f7-8ea5-df3252d70653-config\") pod \"controller-manager-879f6c89f-9z47r\" (UID: \"7a81f8b4-7529-45f7-8ea5-df3252d70653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853356 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spclf\" (UniqueName: \"kubernetes.io/projected/e8d90978-a395-4313-bcb2-ca35a303b518-kube-api-access-spclf\") pod \"cluster-samples-operator-665b6dd947-csgb2\" (UID: \"e8d90978-a395-4313-bcb2-ca35a303b518\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-csgb2" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853375 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/887c5808-d4ac-4517-a1b9-1601f79aafa8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-gpr4j\" (UID: \"887c5808-d4ac-4517-a1b9-1601f79aafa8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gpr4j" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853395 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853412 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2f79e15-fea6-4391-a711-53ca225b1857-trusted-ca\") pod \"ingress-operator-5b745b69d9-pbpvq\" (UID: \"c2f79e15-fea6-4391-a711-53ca225b1857\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pbpvq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853432 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2158024-4bf6-457d-809d-c389b8ff2309-serving-cert\") pod \"route-controller-manager-6576b87f9c-ssk8b\" (UID: \"b2158024-4bf6-457d-809d-c389b8ff2309\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853449 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-image-import-ca\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853467 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/76cb8882-50e3-4443-b73e-27dbb83e4173-proxy-tls\") pod \"machine-config-operator-74547568cd-qm75h\" (UID: \"76cb8882-50e3-4443-b73e-27dbb83e4173\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qm75h" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853487 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-serving-cert\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853508 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g7gm\" (UniqueName: \"kubernetes.io/projected/44a7f769-9a31-4555-be16-51e310ac34e3-kube-api-access-8g7gm\") pod \"downloads-7954f5f757-2kbsj\" (UID: \"44a7f769-9a31-4555-be16-51e310ac34e3\") " pod="openshift-console/downloads-7954f5f757-2kbsj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853528 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c96407ea-0763-479d-8020-ace1607bd427-config\") pod \"authentication-operator-69f744f599-lcljr\" (UID: \"c96407ea-0763-479d-8020-ace1607bd427\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lcljr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853548 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bcca1906-64bf-44dc-9307-b8d44d2b506f-metrics-tls\") pod \"dns-operator-744455d44c-9cnpf\" (UID: \"bcca1906-64bf-44dc-9307-b8d44d2b506f\") " pod="openshift-dns-operator/dns-operator-744455d44c-9cnpf" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853575 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-audit-policies\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853596 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853613 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2158024-4bf6-457d-809d-c389b8ff2309-config\") pod \"route-controller-manager-6576b87f9c-ssk8b\" (UID: \"b2158024-4bf6-457d-809d-c389b8ff2309\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853635 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-audit-policies\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853652 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c96407ea-0763-479d-8020-ace1607bd427-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-lcljr\" (UID: \"c96407ea-0763-479d-8020-ace1607bd427\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lcljr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853670 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-config\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853687 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c2f79e15-fea6-4391-a711-53ca225b1857-metrics-tls\") pod \"ingress-operator-5b745b69d9-pbpvq\" (UID: \"c2f79e15-fea6-4391-a711-53ca225b1857\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pbpvq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853703 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b57dcc9c-6002-4c47-ae85-c287659cb679-etcd-client\") pod \"etcd-operator-b45778765-l74cd\" (UID: \"b57dcc9c-6002-4c47-ae85-c287659cb679\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853721 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a185671e-0a3a-4a9e-a884-65b448b1e922-images\") pod \"machine-api-operator-5694c8668f-j6kt6\" (UID: \"a185671e-0a3a-4a9e-a884-65b448b1e922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6kt6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853741 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9whn\" (UniqueName: \"kubernetes.io/projected/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-kube-api-access-x9whn\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853757 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a81f8b4-7529-45f7-8ea5-df3252d70653-serving-cert\") pod \"controller-manager-879f6c89f-9z47r\" (UID: \"7a81f8b4-7529-45f7-8ea5-df3252d70653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853775 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/16484304-6976-45e8-a495-f1d2ee367f0f-console-serving-cert\") pod \"console-f9d7485db-gx6qj\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853791 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25a48963-29f5-4d84-91cf-b80e380ed9f1-config\") pod \"kube-controller-manager-operator-78b949d7b-pk75f\" (UID: \"25a48963-29f5-4d84-91cf-b80e380ed9f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pk75f" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853807 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxr6m\" (UniqueName: \"kubernetes.io/projected/a185671e-0a3a-4a9e-a884-65b448b1e922-kube-api-access-jxr6m\") pod \"machine-api-operator-5694c8668f-j6kt6\" (UID: \"a185671e-0a3a-4a9e-a884-65b448b1e922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6kt6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853835 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/83d9c4d8-10b5-4d1e-a71e-e6e92efd8379-trusted-ca\") pod \"console-operator-58897d9998-8ks9s\" (UID: \"83d9c4d8-10b5-4d1e-a71e-e6e92efd8379\") " pod="openshift-console-operator/console-operator-58897d9998-8ks9s" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853852 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853869 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853886 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-serving-cert\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853905 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853939 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853960 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83d9c4d8-10b5-4d1e-a71e-e6e92efd8379-config\") pod \"console-operator-58897d9998-8ks9s\" (UID: \"83d9c4d8-10b5-4d1e-a71e-e6e92efd8379\") " pod="openshift-console-operator/console-operator-58897d9998-8ks9s" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.853978 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gztm5\" (UniqueName: \"kubernetes.io/projected/b9f835b0-16cf-4f05-a8a2-2767a929ca25-kube-api-access-gztm5\") pod \"multus-admission-controller-857f4d67dd-rx99b\" (UID: \"b9f835b0-16cf-4f05-a8a2-2767a929ca25\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rx99b" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854000 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0fe547bf-ea91-4546-a886-c613abeae02a-service-ca-bundle\") pod \"router-default-5444994796-r96rq\" (UID: \"0fe547bf-ea91-4546-a886-c613abeae02a\") " pod="openshift-ingress/router-default-5444994796-r96rq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854018 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7826m\" (UniqueName: \"kubernetes.io/projected/bcca1906-64bf-44dc-9307-b8d44d2b506f-kube-api-access-7826m\") pod \"dns-operator-744455d44c-9cnpf\" (UID: \"bcca1906-64bf-44dc-9307-b8d44d2b506f\") " pod="openshift-dns-operator/dns-operator-744455d44c-9cnpf" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854044 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-audit-dir\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854061 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b9f835b0-16cf-4f05-a8a2-2767a929ca25-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rx99b\" (UID: \"b9f835b0-16cf-4f05-a8a2-2767a929ca25\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rx99b" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854080 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-service-ca\") pod \"console-f9d7485db-gx6qj\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854098 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-trusted-ca-bundle\") pod \"console-f9d7485db-gx6qj\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854117 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-oauth-serving-cert\") pod \"console-f9d7485db-gx6qj\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854136 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/25a48963-29f5-4d84-91cf-b80e380ed9f1-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pk75f\" (UID: \"25a48963-29f5-4d84-91cf-b80e380ed9f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pk75f" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854155 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a185671e-0a3a-4a9e-a884-65b448b1e922-config\") pod \"machine-api-operator-5694c8668f-j6kt6\" (UID: \"a185671e-0a3a-4a9e-a884-65b448b1e922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6kt6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854173 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854191 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b57dcc9c-6002-4c47-ae85-c287659cb679-config\") pod \"etcd-operator-b45778765-l74cd\" (UID: \"b57dcc9c-6002-4c47-ae85-c287659cb679\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854210 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/16484304-6976-45e8-a495-f1d2ee367f0f-console-oauth-config\") pod \"console-f9d7485db-gx6qj\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854229 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c96407ea-0763-479d-8020-ace1607bd427-service-ca-bundle\") pod \"authentication-operator-69f744f599-lcljr\" (UID: \"c96407ea-0763-479d-8020-ace1607bd427\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lcljr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854246 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xzss\" (UniqueName: \"kubernetes.io/projected/3937dccc-95e5-4aa3-be81-6d12c31f7cb9-kube-api-access-2xzss\") pod \"openshift-controller-manager-operator-756b6f6bc6-6v84k\" (UID: \"3937dccc-95e5-4aa3-be81-6d12c31f7cb9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6v84k" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854265 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/887c5808-d4ac-4517-a1b9-1601f79aafa8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-gpr4j\" (UID: \"887c5808-d4ac-4517-a1b9-1601f79aafa8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gpr4j" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854285 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/76cb8882-50e3-4443-b73e-27dbb83e4173-images\") pod \"machine-config-operator-74547568cd-qm75h\" (UID: \"76cb8882-50e3-4443-b73e-27dbb83e4173\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qm75h" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854304 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-etcd-client\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854320 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7a81f8b4-7529-45f7-8ea5-df3252d70653-client-ca\") pod \"controller-manager-879f6c89f-9z47r\" (UID: \"7a81f8b4-7529-45f7-8ea5-df3252d70653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854338 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jcb9\" (UniqueName: \"kubernetes.io/projected/0fe547bf-ea91-4546-a886-c613abeae02a-kube-api-access-4jcb9\") pod \"router-default-5444994796-r96rq\" (UID: \"0fe547bf-ea91-4546-a886-c613abeae02a\") " pod="openshift-ingress/router-default-5444994796-r96rq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854360 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5072c69e-c292-478d-9998-8b5237dfc33f-audit-dir\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854377 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854396 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpqnp\" (UniqueName: \"kubernetes.io/projected/b2158024-4bf6-457d-809d-c389b8ff2309-kube-api-access-mpqnp\") pod \"route-controller-manager-6576b87f9c-ssk8b\" (UID: \"b2158024-4bf6-457d-809d-c389b8ff2309\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854415 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e8d90978-a395-4313-bcb2-ca35a303b518-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-csgb2\" (UID: \"e8d90978-a395-4313-bcb2-ca35a303b518\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-csgb2" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854437 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854459 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v5tw\" (UniqueName: \"kubernetes.io/projected/c96407ea-0763-479d-8020-ace1607bd427-kube-api-access-5v5tw\") pod \"authentication-operator-69f744f599-lcljr\" (UID: \"c96407ea-0763-479d-8020-ace1607bd427\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lcljr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854483 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.854500 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-console-config\") pod \"console-f9d7485db-gx6qj\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.855274 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-console-config\") pod \"console-f9d7485db-gx6qj\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.855365 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-audit-dir\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.855961 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/b57dcc9c-6002-4c47-ae85-c287659cb679-etcd-service-ca\") pod \"etcd-operator-b45778765-l74cd\" (UID: \"b57dcc9c-6002-4c47-ae85-c287659cb679\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.856516 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-etcd-serving-ca\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.856903 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/b57dcc9c-6002-4c47-ae85-c287659cb679-etcd-ca\") pod \"etcd-operator-b45778765-l74cd\" (UID: \"b57dcc9c-6002-4c47-ae85-c287659cb679\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.859665 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.859841 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.860537 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-service-ca\") pod \"console-f9d7485db-gx6qj\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.863988 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xd6tf"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.864890 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6v84k"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.864939 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-54dh8"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.865039 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xd6tf" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.865095 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c96407ea-0763-479d-8020-ace1607bd427-serving-cert\") pod \"authentication-operator-69f744f599-lcljr\" (UID: \"c96407ea-0763-479d-8020-ace1607bd427\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lcljr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.865684 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-audit-dir\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.868473 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a185671e-0a3a-4a9e-a884-65b448b1e922-images\") pod \"machine-api-operator-5694c8668f-j6kt6\" (UID: \"a185671e-0a3a-4a9e-a884-65b448b1e922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6kt6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.869161 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-config\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.869648 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.870410 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a185671e-0a3a-4a9e-a884-65b448b1e922-config\") pod \"machine-api-operator-5694c8668f-j6kt6\" (UID: \"a185671e-0a3a-4a9e-a884-65b448b1e922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6kt6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.870889 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.871323 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83d9c4d8-10b5-4d1e-a71e-e6e92efd8379-config\") pod \"console-operator-58897d9998-8ks9s\" (UID: \"83d9c4d8-10b5-4d1e-a71e-e6e92efd8379\") " pod="openshift-console-operator/console-operator-58897d9998-8ks9s" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.871490 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-oauth-serving-cert\") pod \"console-f9d7485db-gx6qj\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.874118 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b57dcc9c-6002-4c47-ae85-c287659cb679-etcd-client\") pod \"etcd-operator-b45778765-l74cd\" (UID: \"b57dcc9c-6002-4c47-ae85-c287659cb679\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.879967 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9z47r"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.881793 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.883740 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-node-pullsecrets\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.885589 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3937dccc-95e5-4aa3-be81-6d12c31f7cb9-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-6v84k\" (UID: \"3937dccc-95e5-4aa3-be81-6d12c31f7cb9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6v84k" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.889733 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c96407ea-0763-479d-8020-ace1607bd427-config\") pod \"authentication-operator-69f744f599-lcljr\" (UID: \"c96407ea-0763-479d-8020-ace1607bd427\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lcljr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.890330 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-audit-policies\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.890605 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-audit-policies\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.892193 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25a48963-29f5-4d84-91cf-b80e380ed9f1-config\") pod \"kube-controller-manager-operator-78b949d7b-pk75f\" (UID: \"25a48963-29f5-4d84-91cf-b80e380ed9f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pk75f" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.892686 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vj8ww"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.892816 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a81f8b4-7529-45f7-8ea5-df3252d70653-serving-cert\") pod \"controller-manager-879f6c89f-9z47r\" (UID: \"7a81f8b4-7529-45f7-8ea5-df3252d70653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.893931 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vj8ww" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.895160 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-serving-cert\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.896633 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2158024-4bf6-457d-809d-c389b8ff2309-config\") pod \"route-controller-manager-6576b87f9c-ssk8b\" (UID: \"b2158024-4bf6-457d-809d-c389b8ff2309\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.898447 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k2lwz"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.901361 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k2lwz" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.902480 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-encryption-config\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.904264 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/76cb8882-50e3-4443-b73e-27dbb83e4173-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qm75h\" (UID: \"76cb8882-50e3-4443-b73e-27dbb83e4173\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qm75h" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.904281 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.919495 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2158024-4bf6-457d-809d-c389b8ff2309-serving-cert\") pod \"route-controller-manager-6576b87f9c-ssk8b\" (UID: \"b2158024-4bf6-457d-809d-c389b8ff2309\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.905459 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-image-import-ca\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.906329 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a81f8b4-7529-45f7-8ea5-df3252d70653-config\") pod \"controller-manager-879f6c89f-9z47r\" (UID: \"7a81f8b4-7529-45f7-8ea5-df3252d70653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.906651 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b57dcc9c-6002-4c47-ae85-c287659cb679-config\") pod \"etcd-operator-b45778765-l74cd\" (UID: \"b57dcc9c-6002-4c47-ae85-c287659cb679\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.907426 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.904882 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3937dccc-95e5-4aa3-be81-6d12c31f7cb9-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-6v84k\" (UID: \"3937dccc-95e5-4aa3-be81-6d12c31f7cb9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6v84k" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.920168 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c96407ea-0763-479d-8020-ace1607bd427-service-ca-bundle\") pod \"authentication-operator-69f744f599-lcljr\" (UID: \"c96407ea-0763-479d-8020-ace1607bd427\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lcljr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.920522 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/16484304-6976-45e8-a495-f1d2ee367f0f-console-serving-cert\") pod \"console-f9d7485db-gx6qj\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.920992 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.921599 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.951161 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.962138 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c2f79e15-fea6-4391-a711-53ca225b1857-metrics-tls\") pod \"ingress-operator-5b745b69d9-pbpvq\" (UID: \"c2f79e15-fea6-4391-a711-53ca225b1857\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pbpvq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.962532 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.962890 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-serving-cert\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.967429 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-encryption-config\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.967884 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25a48963-29f5-4d84-91cf-b80e380ed9f1-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pk75f\" (UID: \"25a48963-29f5-4d84-91cf-b80e380ed9f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pk75f" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.968203 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.951520 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.970665 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.952477 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.953360 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.968534 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.971970 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.973832 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-trusted-ca-bundle\") pod \"console-f9d7485db-gx6qj\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.974933 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83d9c4d8-10b5-4d1e-a71e-e6e92efd8379-serving-cert\") pod \"console-operator-58897d9998-8ks9s\" (UID: \"83d9c4d8-10b5-4d1e-a71e-e6e92efd8379\") " pod="openshift-console-operator/console-operator-58897d9998-8ks9s" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.975582 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.983555 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.984164 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.984839 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.985022 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.985135 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5072c69e-c292-478d-9998-8b5237dfc33f-audit-dir\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.985687 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.985788 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b2158024-4bf6-457d-809d-c389b8ff2309-client-ca\") pod \"route-controller-manager-6576b87f9c-ssk8b\" (UID: \"b2158024-4bf6-457d-809d-c389b8ff2309\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.986251 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.986500 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.989256 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c96407ea-0763-479d-8020-ace1607bd427-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-lcljr\" (UID: \"c96407ea-0763-479d-8020-ace1607bd427\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lcljr" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.989676 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.989824 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.990109 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.990126 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-etcd-client\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.990270 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.990899 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2f79e15-fea6-4391-a711-53ca225b1857-trusted-ca\") pod \"ingress-operator-5b745b69d9-pbpvq\" (UID: \"c2f79e15-fea6-4391-a711-53ca225b1857\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pbpvq" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.991528 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e8d90978-a395-4313-bcb2-ca35a303b518-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-csgb2\" (UID: \"e8d90978-a395-4313-bcb2-ca35a303b518\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-csgb2" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.991792 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/16484304-6976-45e8-a495-f1d2ee367f0f-console-oauth-config\") pod \"console-f9d7485db-gx6qj\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.991859 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7a81f8b4-7529-45f7-8ea5-df3252d70653-client-ca\") pod \"controller-manager-879f6c89f-9z47r\" (UID: \"7a81f8b4-7529-45f7-8ea5-df3252d70653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.993183 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.993316 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.994662 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.995786 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b57dcc9c-6002-4c47-ae85-c287659cb679-serving-cert\") pod \"etcd-operator-b45778765-l74cd\" (UID: \"b57dcc9c-6002-4c47-ae85-c287659cb679\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.995812 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a185671e-0a3a-4a9e-a884-65b448b1e922-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j6kt6\" (UID: \"a185671e-0a3a-4a9e-a884-65b448b1e922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6kt6" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.996115 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.997159 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.997293 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kgrk6"] Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.997904 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/887c5808-d4ac-4517-a1b9-1601f79aafa8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-gpr4j\" (UID: \"887c5808-d4ac-4517-a1b9-1601f79aafa8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gpr4j" Oct 13 08:46:53 crc kubenswrapper[4685]: I1013 08:46:53.999298 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-etcd-client\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:53.999981 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bcca1906-64bf-44dc-9307-b8d44d2b506f-metrics-tls\") pod \"dns-operator-744455d44c-9cnpf\" (UID: \"bcca1906-64bf-44dc-9307-b8d44d2b506f\") " pod="openshift-dns-operator/dns-operator-744455d44c-9cnpf" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.000606 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7a81f8b4-7529-45f7-8ea5-df3252d70653-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9z47r\" (UID: \"7a81f8b4-7529-45f7-8ea5-df3252d70653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.000742 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4zvdw"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.001841 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.002305 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.002637 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4zvdw" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.004473 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.003582 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/83d9c4d8-10b5-4d1e-a71e-e6e92efd8379-trusted-ca\") pod \"console-operator-58897d9998-8ks9s\" (UID: \"83d9c4d8-10b5-4d1e-a71e-e6e92efd8379\") " pod="openshift-console-operator/console-operator-58897d9998-8ks9s" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.008068 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7nbg"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.008901 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7nbg" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.010819 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7sx5f"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.011411 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.013462 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9wxlv"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.015031 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.015356 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2bwvx"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.016004 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-9wxlv" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.018566 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.018948 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2bwvx" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.020227 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.020331 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8ks9s"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.020397 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-2kbsj"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.020335 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.021682 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-l74cd"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.022835 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-gx6qj"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.024356 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-lcljr"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.025394 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pk75f"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.026553 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-pbpvq"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.028153 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wn86t"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.035125 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qm75h"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.035165 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8jzsj"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.035179 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-csgb2"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.035197 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j6kt6"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.035317 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn86t" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.035344 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84w94"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.035650 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.036838 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-wmjcx"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.037690 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-9cnpf"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.037815 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-wmjcx" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.038720 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-96w58"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.039774 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-96w58" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.041247 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/887c5808-d4ac-4517-a1b9-1601f79aafa8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-gpr4j\" (UID: \"887c5808-d4ac-4517-a1b9-1601f79aafa8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gpr4j" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.044978 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-n8hmz"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.046087 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rx99b"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.047568 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k2lwz"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.049290 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.051254 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9wxlv"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.052467 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7nbg"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.053682 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-cnh25"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.060985 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-cnh25" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.061935 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8x7bf"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.065895 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xd6tf"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.068319 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2bwvx"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.070661 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.070890 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gpr4j"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.073814 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ljhpv"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.074355 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.075898 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7sx5f"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.077438 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-4lflr"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.078955 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-cnh25"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.080667 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vj8ww"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.082187 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-ps67b"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.082983 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-ps67b" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.084218 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-96w58"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.085346 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.086427 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4zvdw"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.087677 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wn86t"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.088642 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-ps67b"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.094460 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.115092 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.135497 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.155169 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.174789 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.181632 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0fe547bf-ea91-4546-a886-c613abeae02a-stats-auth\") pod \"router-default-5444994796-r96rq\" (UID: \"0fe547bf-ea91-4546-a886-c613abeae02a\") " pod="openshift-ingress/router-default-5444994796-r96rq" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.195270 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.215532 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.216266 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-audit\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.227155 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0fe547bf-ea91-4546-a886-c613abeae02a-metrics-certs\") pod \"router-default-5444994796-r96rq\" (UID: \"0fe547bf-ea91-4546-a886-c613abeae02a\") " pod="openshift-ingress/router-default-5444994796-r96rq" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.235052 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.239340 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0fe547bf-ea91-4546-a886-c613abeae02a-default-certificate\") pod \"router-default-5444994796-r96rq\" (UID: \"0fe547bf-ea91-4546-a886-c613abeae02a\") " pod="openshift-ingress/router-default-5444994796-r96rq" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.255230 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.275388 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.279435 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0fe547bf-ea91-4546-a886-c613abeae02a-service-ca-bundle\") pod \"router-default-5444994796-r96rq\" (UID: \"0fe547bf-ea91-4546-a886-c613abeae02a\") " pod="openshift-ingress/router-default-5444994796-r96rq" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.295935 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.315755 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.324885 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/76cb8882-50e3-4443-b73e-27dbb83e4173-proxy-tls\") pod \"machine-config-operator-74547568cd-qm75h\" (UID: \"76cb8882-50e3-4443-b73e-27dbb83e4173\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qm75h" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.334945 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.338130 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b9f835b0-16cf-4f05-a8a2-2767a929ca25-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rx99b\" (UID: \"b9f835b0-16cf-4f05-a8a2-2767a929ca25\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rx99b" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.355480 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.359420 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/76cb8882-50e3-4443-b73e-27dbb83e4173-images\") pod \"machine-config-operator-74547568cd-qm75h\" (UID: \"76cb8882-50e3-4443-b73e-27dbb83e4173\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qm75h" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.375230 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.415411 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.434331 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.454258 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.475515 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.497011 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.515826 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.535558 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.556146 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.576254 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.595459 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.630721 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbp54\" (UniqueName: \"kubernetes.io/projected/8f247d85-48fa-49a1-9c0c-3ec7b7d202c0-kube-api-access-nbp54\") pod \"apiserver-76f77b778f-kgrk6\" (UID: \"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0\") " pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.649683 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z579t\" (UniqueName: \"kubernetes.io/projected/76cb8882-50e3-4443-b73e-27dbb83e4173-kube-api-access-z579t\") pod \"machine-config-operator-74547568cd-qm75h\" (UID: \"76cb8882-50e3-4443-b73e-27dbb83e4173\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qm75h" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.665521 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qm75h" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.672633 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl6bm\" (UniqueName: \"kubernetes.io/projected/b57dcc9c-6002-4c47-ae85-c287659cb679-kube-api-access-pl6bm\") pod \"etcd-operator-b45778765-l74cd\" (UID: \"b57dcc9c-6002-4c47-ae85-c287659cb679\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.690613 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k2b5\" (UniqueName: \"kubernetes.io/projected/5072c69e-c292-478d-9998-8b5237dfc33f-kube-api-access-5k2b5\") pod \"oauth-openshift-558db77b4-54dh8\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.695187 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.730850 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gztm5\" (UniqueName: \"kubernetes.io/projected/b9f835b0-16cf-4f05-a8a2-2767a929ca25-kube-api-access-gztm5\") pod \"multus-admission-controller-857f4d67dd-rx99b\" (UID: \"b9f835b0-16cf-4f05-a8a2-2767a929ca25\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rx99b" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.735820 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.751853 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.766901 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.772308 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7826m\" (UniqueName: \"kubernetes.io/projected/bcca1906-64bf-44dc-9307-b8d44d2b506f-kube-api-access-7826m\") pod \"dns-operator-744455d44c-9cnpf\" (UID: \"bcca1906-64bf-44dc-9307-b8d44d2b506f\") " pod="openshift-dns-operator/dns-operator-744455d44c-9cnpf" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.797789 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c2f79e15-fea6-4391-a711-53ca225b1857-bound-sa-token\") pod \"ingress-operator-5b745b69d9-pbpvq\" (UID: \"c2f79e15-fea6-4391-a711-53ca225b1857\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pbpvq" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.815827 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh6nw\" (UniqueName: \"kubernetes.io/projected/7a81f8b4-7529-45f7-8ea5-df3252d70653-kube-api-access-nh6nw\") pod \"controller-manager-879f6c89f-9z47r\" (UID: \"7a81f8b4-7529-45f7-8ea5-df3252d70653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.829856 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/25a48963-29f5-4d84-91cf-b80e380ed9f1-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pk75f\" (UID: \"25a48963-29f5-4d84-91cf-b80e380ed9f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pk75f" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.835882 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.836731 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.858699 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.875569 4685 request.go:700] Waited for 1.007862888s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console/serviceaccounts/console/token Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.875613 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.893647 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hfd8\" (UniqueName: \"kubernetes.io/projected/16484304-6976-45e8-a495-f1d2ee367f0f-kube-api-access-6hfd8\") pod \"console-f9d7485db-gx6qj\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.893873 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pk75f" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.896450 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.917701 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-9cnpf" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.932191 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.939086 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9whn\" (UniqueName: \"kubernetes.io/projected/2f11d8d5-7c31-42a1-83e5-cfbfa7de648d-kube-api-access-x9whn\") pod \"apiserver-7bbb656c7d-f46fr\" (UID: \"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.940070 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qm75h"] Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.975494 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-rx99b" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.983179 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgvfj\" (UniqueName: \"kubernetes.io/projected/887c5808-d4ac-4517-a1b9-1601f79aafa8-kube-api-access-jgvfj\") pod \"openshift-apiserver-operator-796bbdcf4f-gpr4j\" (UID: \"887c5808-d4ac-4517-a1b9-1601f79aafa8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gpr4j" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.988495 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g7gm\" (UniqueName: \"kubernetes.io/projected/44a7f769-9a31-4555-be16-51e310ac34e3-kube-api-access-8g7gm\") pod \"downloads-7954f5f757-2kbsj\" (UID: \"44a7f769-9a31-4555-be16-51e310ac34e3\") " pod="openshift-console/downloads-7954f5f757-2kbsj" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.997358 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 13 08:46:54 crc kubenswrapper[4685]: I1013 08:46:54.997634 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxr6m\" (UniqueName: \"kubernetes.io/projected/a185671e-0a3a-4a9e-a884-65b448b1e922-kube-api-access-jxr6m\") pod \"machine-api-operator-5694c8668f-j6kt6\" (UID: \"a185671e-0a3a-4a9e-a884-65b448b1e922\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j6kt6" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.019193 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.052662 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spclf\" (UniqueName: \"kubernetes.io/projected/e8d90978-a395-4313-bcb2-ca35a303b518-kube-api-access-spclf\") pod \"cluster-samples-operator-665b6dd947-csgb2\" (UID: \"e8d90978-a395-4313-bcb2-ca35a303b518\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-csgb2" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.080227 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.126967 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfltc\" (UniqueName: \"kubernetes.io/projected/c2f79e15-fea6-4391-a711-53ca225b1857-kube-api-access-vfltc\") pod \"ingress-operator-5b745b69d9-pbpvq\" (UID: \"c2f79e15-fea6-4391-a711-53ca225b1857\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pbpvq" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.135936 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.140512 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s4dp\" (UniqueName: \"kubernetes.io/projected/83d9c4d8-10b5-4d1e-a71e-e6e92efd8379-kube-api-access-9s4dp\") pod \"console-operator-58897d9998-8ks9s\" (UID: \"83d9c4d8-10b5-4d1e-a71e-e6e92efd8379\") " pod="openshift-console-operator/console-operator-58897d9998-8ks9s" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.158699 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.158762 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kgrk6"] Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.160367 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.167776 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-54dh8"] Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.188896 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xzss\" (UniqueName: \"kubernetes.io/projected/3937dccc-95e5-4aa3-be81-6d12c31f7cb9-kube-api-access-2xzss\") pod \"openshift-controller-manager-operator-756b6f6bc6-6v84k\" (UID: \"3937dccc-95e5-4aa3-be81-6d12c31f7cb9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6v84k" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.193711 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-csgb2" Oct 13 08:46:55 crc kubenswrapper[4685]: W1013 08:46:55.205409 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5072c69e_c292_478d_9998_8b5237dfc33f.slice/crio-ce543441c286144011a8d80fdbf12acb00250210fa9082c11b50b6d1ff64be0d WatchSource:0}: Error finding container ce543441c286144011a8d80fdbf12acb00250210fa9082c11b50b6d1ff64be0d: Status 404 returned error can't find the container with id ce543441c286144011a8d80fdbf12acb00250210fa9082c11b50b6d1ff64be0d Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.206033 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6v84k" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.216789 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-8ks9s" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.230160 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpqnp\" (UniqueName: \"kubernetes.io/projected/b2158024-4bf6-457d-809d-c389b8ff2309-kube-api-access-mpqnp\") pod \"route-controller-manager-6576b87f9c-ssk8b\" (UID: \"b2158024-4bf6-457d-809d-c389b8ff2309\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.231458 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pk75f"] Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.232298 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gpr4j" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.240263 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v5tw\" (UniqueName: \"kubernetes.io/projected/c96407ea-0763-479d-8020-ace1607bd427-kube-api-access-5v5tw\") pod \"authentication-operator-69f744f599-lcljr\" (UID: \"c96407ea-0763-479d-8020-ace1607bd427\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lcljr" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.247404 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-2kbsj" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.255546 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.257544 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jcb9\" (UniqueName: \"kubernetes.io/projected/0fe547bf-ea91-4546-a886-c613abeae02a-kube-api-access-4jcb9\") pod \"router-default-5444994796-r96rq\" (UID: \"0fe547bf-ea91-4546-a886-c613abeae02a\") " pod="openshift-ingress/router-default-5444994796-r96rq" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.261812 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" event={"ID":"5072c69e-c292-478d-9998-8b5237dfc33f","Type":"ContainerStarted","Data":"ce543441c286144011a8d80fdbf12acb00250210fa9082c11b50b6d1ff64be0d"} Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.266867 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qm75h" event={"ID":"76cb8882-50e3-4443-b73e-27dbb83e4173","Type":"ContainerStarted","Data":"e6b9b048a688b2f7065d98ec19ea3e7afb1bb3b3da8af4b58a55c5d64c0c80f4"} Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.267033 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qm75h" event={"ID":"76cb8882-50e3-4443-b73e-27dbb83e4173","Type":"ContainerStarted","Data":"6f654e28c904e37130dec31ff6c60392ed40ed44b3cdb515540457ab38516a4e"} Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.267668 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" event={"ID":"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0","Type":"ContainerStarted","Data":"221f7ab0cf343526a302dcba6234149f998dcc63ffb9b7ad611edff0ec5c5144"} Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.275725 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-j6kt6" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.278849 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.279684 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-l74cd"] Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.286537 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pbpvq" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.296532 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.315871 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.331279 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9z47r"] Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.338285 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.357344 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 13 08:46:55 crc kubenswrapper[4685]: W1013 08:46:55.373539 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb57dcc9c_6002_4c47_ae85_c287659cb679.slice/crio-4de31c7db49bd4a2134c1bb5fc7d484724396de62d0523094f46f6411ef74b79 WatchSource:0}: Error finding container 4de31c7db49bd4a2134c1bb5fc7d484724396de62d0523094f46f6411ef74b79: Status 404 returned error can't find the container with id 4de31c7db49bd4a2134c1bb5fc7d484724396de62d0523094f46f6411ef74b79 Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.375866 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 13 08:46:55 crc kubenswrapper[4685]: W1013 08:46:55.385426 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a81f8b4_7529_45f7_8ea5_df3252d70653.slice/crio-20b415acc33a00de655d765eec169b0519ed69c6923a8371aac4eb82f8cf502a WatchSource:0}: Error finding container 20b415acc33a00de655d765eec169b0519ed69c6923a8371aac4eb82f8cf502a: Status 404 returned error can't find the container with id 20b415acc33a00de655d765eec169b0519ed69c6923a8371aac4eb82f8cf502a Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.406715 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.415463 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-gx6qj"] Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.419222 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.434865 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.456078 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.461142 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-csgb2"] Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.465362 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.479258 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.483864 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-lcljr" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.496391 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.517324 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.540962 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.551357 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-r96rq" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.556354 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.592352 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.595882 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.615483 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.634970 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.655712 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.656370 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rx99b"] Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.656400 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-9cnpf"] Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.656410 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6v84k"] Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.656419 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr"] Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.676040 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.707733 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.717848 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.736338 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.758376 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.776205 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.795324 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.816271 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.819649 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gpr4j"] Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.838320 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.845904 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8ks9s"] Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.856283 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.876286 4685 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.894170 4685 request.go:700] Waited for 1.832868854s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/configmaps?fieldSelector=metadata.name%3Ddns-default&limit=500&resourceVersion=0 Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.896382 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.917876 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.941281 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 13 08:46:55 crc kubenswrapper[4685]: W1013 08:46:55.953465 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0fe547bf_ea91_4546_a886_c613abeae02a.slice/crio-f65a4507bc119d147d83f5add56a3c2e8c0eaa0ffdb2cdb69a4aec8117465612 WatchSource:0}: Error finding container f65a4507bc119d147d83f5add56a3c2e8c0eaa0ffdb2cdb69a4aec8117465612: Status 404 returned error can't find the container with id f65a4507bc119d147d83f5add56a3c2e8c0eaa0ffdb2cdb69a4aec8117465612 Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.966128 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.979531 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 13 08:46:55 crc kubenswrapper[4685]: I1013 08:46:55.996337 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.015707 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.134034 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/70fe2241-a755-4556-a84a-733306f90d66-registry-certificates\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.134137 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-84w94\" (UID: \"30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84w94" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.134337 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/70fe2241-a755-4556-a84a-733306f90d66-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.134423 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kz5ht\" (UniqueName: \"kubernetes.io/projected/70fe2241-a755-4556-a84a-733306f90d66-kube-api-access-kz5ht\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.134444 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa9441a2-a8c5-4f10-89c6-88fa689e9f7c-serving-cert\") pod \"openshift-config-operator-7777fb866f-4lflr\" (UID: \"aa9441a2-a8c5-4f10-89c6-88fa689e9f7c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4lflr" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.134490 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/70fe2241-a755-4556-a84a-733306f90d66-registry-tls\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.134604 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/70fe2241-a755-4556-a84a-733306f90d66-bound-sa-token\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.134628 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6e5e0a0-6ff8-44eb-99f6-eb9a4eb2aa47-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8x7bf\" (UID: \"e6e5e0a0-6ff8-44eb-99f6-eb9a4eb2aa47\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8x7bf" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.134663 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/70fe2241-a755-4556-a84a-733306f90d66-trusted-ca\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.134699 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6e5e0a0-6ff8-44eb-99f6-eb9a4eb2aa47-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8x7bf\" (UID: \"e6e5e0a0-6ff8-44eb-99f6-eb9a4eb2aa47\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8x7bf" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.134758 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw8rq\" (UniqueName: \"kubernetes.io/projected/aa9441a2-a8c5-4f10-89c6-88fa689e9f7c-kube-api-access-vw8rq\") pod \"openshift-config-operator-7777fb866f-4lflr\" (UID: \"aa9441a2-a8c5-4f10-89c6-88fa689e9f7c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4lflr" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.138593 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-2kbsj"] Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.139062 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-84w94\" (UID: \"30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84w94" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.139483 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.139595 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/70fe2241-a755-4556-a84a-733306f90d66-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: E1013 08:46:56.139890 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:46:56.639877624 +0000 UTC m=+141.787753385 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.141257 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvqmj\" (UniqueName: \"kubernetes.io/projected/30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90-kube-api-access-jvqmj\") pod \"cluster-image-registry-operator-dc59b4c8b-84w94\" (UID: \"30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84w94" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.141647 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-84w94\" (UID: \"30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84w94" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.141854 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmx29\" (UniqueName: \"kubernetes.io/projected/e6e5e0a0-6ff8-44eb-99f6-eb9a4eb2aa47-kube-api-access-fmx29\") pod \"kube-storage-version-migrator-operator-b67b599dd-8x7bf\" (UID: \"e6e5e0a0-6ff8-44eb-99f6-eb9a4eb2aa47\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8x7bf" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.142175 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/aa9441a2-a8c5-4f10-89c6-88fa689e9f7c-available-featuregates\") pod \"openshift-config-operator-7777fb866f-4lflr\" (UID: \"aa9441a2-a8c5-4f10-89c6-88fa689e9f7c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4lflr" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.171000 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-pbpvq"] Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.171471 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j6kt6"] Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.190181 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b"] Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.244872 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.245297 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8p545\" (UniqueName: \"kubernetes.io/projected/c1eea5fd-9b4d-4d4d-a715-6523309cdd87-kube-api-access-8p545\") pod \"catalog-operator-68c6474976-vj8ww\" (UID: \"c1eea5fd-9b4d-4d4d-a715-6523309cdd87\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vj8ww" Oct 13 08:46:56 crc kubenswrapper[4685]: E1013 08:46:56.245422 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:46:56.745378317 +0000 UTC m=+141.893254078 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.245531 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ef87f11-a2ee-48e5-86ec-94283d235933-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2bwvx\" (UID: \"4ef87f11-a2ee-48e5-86ec-94283d235933\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2bwvx" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.245590 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/d869fa03-5196-4a23-a7d3-9bb709891678-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-4zvdw\" (UID: \"d869fa03-5196-4a23-a7d3-9bb709891678\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4zvdw" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.245639 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/926501e1-d68a-4c86-b59e-f87c43c3a9ae-webhook-cert\") pod \"packageserver-d55dfcdfc-vh7kj\" (UID: \"926501e1-d68a-4c86-b59e-f87c43c3a9ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.245664 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa9441a2-a8c5-4f10-89c6-88fa689e9f7c-serving-cert\") pod \"openshift-config-operator-7777fb866f-4lflr\" (UID: \"aa9441a2-a8c5-4f10-89c6-88fa689e9f7c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4lflr" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.245686 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/53a5a736-ebd0-4c09-90b3-d70a4943c038-config-volume\") pod \"dns-default-cnh25\" (UID: \"53a5a736-ebd0-4c09-90b3-d70a4943c038\") " pod="openshift-dns/dns-default-cnh25" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.245708 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/571d990f-04d9-4e62-902b-f44b7b666901-srv-cert\") pod \"olm-operator-6b444d44fb-k2lwz\" (UID: \"571d990f-04d9-4e62-902b-f44b7b666901\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k2lwz" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.245730 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/763388db-ca9e-432e-9c91-4821392af1a1-plugins-dir\") pod \"csi-hostpathplugin-96w58\" (UID: \"763388db-ca9e-432e-9c91-4821392af1a1\") " pod="hostpath-provisioner/csi-hostpathplugin-96w58" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.245751 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c926cbb-e6ad-4e49-9a6c-afaf6eb9ced6-config\") pod \"kube-apiserver-operator-766d6c64bb-ljhpv\" (UID: \"9c926cbb-e6ad-4e49-9a6c-afaf6eb9ced6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ljhpv" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.245771 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b8ae1ce1-bf5f-4d13-87b4-0810f63434af-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7sx5f\" (UID: \"b8ae1ce1-bf5f-4d13-87b4-0810f63434af\") " pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.245831 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6e5e0a0-6ff8-44eb-99f6-eb9a4eb2aa47-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8x7bf\" (UID: \"e6e5e0a0-6ff8-44eb-99f6-eb9a4eb2aa47\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8x7bf" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.245855 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/763388db-ca9e-432e-9c91-4821392af1a1-socket-dir\") pod \"csi-hostpathplugin-96w58\" (UID: \"763388db-ca9e-432e-9c91-4821392af1a1\") " pod="hostpath-provisioner/csi-hostpathplugin-96w58" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.245877 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/916530ba-091f-4181-aee2-eaa5cd546db5-machine-approver-tls\") pod \"machine-approver-56656f9798-9c4d9\" (UID: \"916530ba-091f-4181-aee2-eaa5cd546db5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c4d9" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.245904 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/70fe2241-a755-4556-a84a-733306f90d66-trusted-ca\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.245944 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6vrw\" (UniqueName: \"kubernetes.io/projected/6e73973b-70de-449a-9ad2-b965b7a83e1f-kube-api-access-n6vrw\") pod \"machine-config-server-wmjcx\" (UID: \"6e73973b-70de-449a-9ad2-b965b7a83e1f\") " pod="openshift-machine-config-operator/machine-config-server-wmjcx" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.245972 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrpk2\" (UniqueName: \"kubernetes.io/projected/3672a087-9d9b-45ee-8f80-686959b395d6-kube-api-access-xrpk2\") pod \"collect-profiles-29339085-84tsn\" (UID: \"3672a087-9d9b-45ee-8f80-686959b395d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.245990 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zbqc\" (UniqueName: \"kubernetes.io/projected/b8ae1ce1-bf5f-4d13-87b4-0810f63434af-kube-api-access-9zbqc\") pod \"marketplace-operator-79b997595-7sx5f\" (UID: \"b8ae1ce1-bf5f-4d13-87b4-0810f63434af\") " pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246069 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6e73973b-70de-449a-9ad2-b965b7a83e1f-node-bootstrap-token\") pod \"machine-config-server-wmjcx\" (UID: \"6e73973b-70de-449a-9ad2-b965b7a83e1f\") " pod="openshift-machine-config-operator/machine-config-server-wmjcx" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246090 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/53a5a736-ebd0-4c09-90b3-d70a4943c038-metrics-tls\") pod \"dns-default-cnh25\" (UID: \"53a5a736-ebd0-4c09-90b3-d70a4943c038\") " pod="openshift-dns/dns-default-cnh25" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246107 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/926501e1-d68a-4c86-b59e-f87c43c3a9ae-tmpfs\") pod \"packageserver-d55dfcdfc-vh7kj\" (UID: \"926501e1-d68a-4c86-b59e-f87c43c3a9ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246149 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246185 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/70fe2241-a755-4556-a84a-733306f90d66-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246206 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvqmj\" (UniqueName: \"kubernetes.io/projected/30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90-kube-api-access-jvqmj\") pod \"cluster-image-registry-operator-dc59b4c8b-84w94\" (UID: \"30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84w94" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246245 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-84w94\" (UID: \"30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84w94" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246270 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztzr7\" (UniqueName: \"kubernetes.io/projected/51d519c2-0060-4aa2-9b25-11c742aca2d1-kube-api-access-ztzr7\") pod \"service-ca-9c57cc56f-9wxlv\" (UID: \"51d519c2-0060-4aa2-9b25-11c742aca2d1\") " pod="openshift-service-ca/service-ca-9c57cc56f-9wxlv" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246310 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/acb21b38-5eb2-4db3-a8ba-9c022c9ae79e-proxy-tls\") pod \"machine-config-controller-84d6567774-n8hmz\" (UID: \"acb21b38-5eb2-4db3-a8ba-9c022c9ae79e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n8hmz" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246383 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b8ae1ce1-bf5f-4d13-87b4-0810f63434af-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7sx5f\" (UID: \"b8ae1ce1-bf5f-4d13-87b4-0810f63434af\") " pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246406 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/aa9441a2-a8c5-4f10-89c6-88fa689e9f7c-available-featuregates\") pod \"openshift-config-operator-7777fb866f-4lflr\" (UID: \"aa9441a2-a8c5-4f10-89c6-88fa689e9f7c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4lflr" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246448 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/763388db-ca9e-432e-9c91-4821392af1a1-mountpoint-dir\") pod \"csi-hostpathplugin-96w58\" (UID: \"763388db-ca9e-432e-9c91-4821392af1a1\") " pod="hostpath-provisioner/csi-hostpathplugin-96w58" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246488 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/926501e1-d68a-4c86-b59e-f87c43c3a9ae-apiservice-cert\") pod \"packageserver-d55dfcdfc-vh7kj\" (UID: \"926501e1-d68a-4c86-b59e-f87c43c3a9ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246542 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7kgm\" (UniqueName: \"kubernetes.io/projected/53a5a736-ebd0-4c09-90b3-d70a4943c038-kube-api-access-v7kgm\") pod \"dns-default-cnh25\" (UID: \"53a5a736-ebd0-4c09-90b3-d70a4943c038\") " pod="openshift-dns/dns-default-cnh25" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246593 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgtzx\" (UniqueName: \"kubernetes.io/projected/acb21b38-5eb2-4db3-a8ba-9c022c9ae79e-kube-api-access-jgtzx\") pod \"machine-config-controller-84d6567774-n8hmz\" (UID: \"acb21b38-5eb2-4db3-a8ba-9c022c9ae79e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n8hmz" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246636 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/916530ba-091f-4181-aee2-eaa5cd546db5-auth-proxy-config\") pod \"machine-approver-56656f9798-9c4d9\" (UID: \"916530ba-091f-4181-aee2-eaa5cd546db5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c4d9" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246667 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c926cbb-e6ad-4e49-9a6c-afaf6eb9ced6-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-ljhpv\" (UID: \"9c926cbb-e6ad-4e49-9a6c-afaf6eb9ced6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ljhpv" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246689 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnldg\" (UniqueName: \"kubernetes.io/projected/47f6f699-d6e2-4289-bd8a-9ec387887523-kube-api-access-tnldg\") pod \"migrator-59844c95c7-xd6tf\" (UID: \"47f6f699-d6e2-4289-bd8a-9ec387887523\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xd6tf" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246714 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ef87f11-a2ee-48e5-86ec-94283d235933-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2bwvx\" (UID: \"4ef87f11-a2ee-48e5-86ec-94283d235933\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2bwvx" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246737 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/70fe2241-a755-4556-a84a-733306f90d66-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246778 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl4cc\" (UniqueName: \"kubernetes.io/projected/d869fa03-5196-4a23-a7d3-9bb709891678-kube-api-access-xl4cc\") pod \"control-plane-machine-set-operator-78cbb6b69f-4zvdw\" (UID: \"d869fa03-5196-4a23-a7d3-9bb709891678\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4zvdw" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246802 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9cjk\" (UniqueName: \"kubernetes.io/projected/9ab91fd8-27d8-458a-97b9-09b9ba008d67-kube-api-access-p9cjk\") pod \"ingress-canary-ps67b\" (UID: \"9ab91fd8-27d8-458a-97b9-09b9ba008d67\") " pod="openshift-ingress-canary/ingress-canary-ps67b" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.246865 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9cdz\" (UniqueName: \"kubernetes.io/projected/bd767915-7aeb-45a4-96a3-c2f636c5e754-kube-api-access-l9cdz\") pod \"package-server-manager-789f6589d5-b7nbg\" (UID: \"bd767915-7aeb-45a4-96a3-c2f636c5e754\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7nbg" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.248148 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/aa9441a2-a8c5-4f10-89c6-88fa689e9f7c-available-featuregates\") pod \"openshift-config-operator-7777fb866f-4lflr\" (UID: \"aa9441a2-a8c5-4f10-89c6-88fa689e9f7c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4lflr" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.249188 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/70fe2241-a755-4556-a84a-733306f90d66-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: E1013 08:46:56.251315 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:46:56.751293393 +0000 UTC m=+141.899169154 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.255847 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6e5e0a0-6ff8-44eb-99f6-eb9a4eb2aa47-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8x7bf\" (UID: \"e6e5e0a0-6ff8-44eb-99f6-eb9a4eb2aa47\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8x7bf" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.256673 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kz5ht\" (UniqueName: \"kubernetes.io/projected/70fe2241-a755-4556-a84a-733306f90d66-kube-api-access-kz5ht\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.256714 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c1eea5fd-9b4d-4d4d-a715-6523309cdd87-srv-cert\") pod \"catalog-operator-68c6474976-vj8ww\" (UID: \"c1eea5fd-9b4d-4d4d-a715-6523309cdd87\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vj8ww" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.256741 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/991a9129-37d1-4407-a9aa-a675512ec8f1-config\") pod \"service-ca-operator-777779d784-wn86t\" (UID: \"991a9129-37d1-4407-a9aa-a675512ec8f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn86t" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.256877 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/70fe2241-a755-4556-a84a-733306f90d66-registry-tls\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.256947 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3672a087-9d9b-45ee-8f80-686959b395d6-config-volume\") pod \"collect-profiles-29339085-84tsn\" (UID: \"3672a087-9d9b-45ee-8f80-686959b395d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.256981 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/70fe2241-a755-4556-a84a-733306f90d66-bound-sa-token\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.257004 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq2tm\" (UniqueName: \"kubernetes.io/projected/916530ba-091f-4181-aee2-eaa5cd546db5-kube-api-access-xq2tm\") pod \"machine-approver-56656f9798-9c4d9\" (UID: \"916530ba-091f-4181-aee2-eaa5cd546db5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c4d9" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.257023 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3672a087-9d9b-45ee-8f80-686959b395d6-secret-volume\") pod \"collect-profiles-29339085-84tsn\" (UID: \"3672a087-9d9b-45ee-8f80-686959b395d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.257472 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/991a9129-37d1-4407-a9aa-a675512ec8f1-serving-cert\") pod \"service-ca-operator-777779d784-wn86t\" (UID: \"991a9129-37d1-4407-a9aa-a675512ec8f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn86t" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.257498 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/51d519c2-0060-4aa2-9b25-11c742aca2d1-signing-key\") pod \"service-ca-9c57cc56f-9wxlv\" (UID: \"51d519c2-0060-4aa2-9b25-11c742aca2d1\") " pod="openshift-service-ca/service-ca-9c57cc56f-9wxlv" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.257519 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhdcr\" (UniqueName: \"kubernetes.io/projected/991a9129-37d1-4407-a9aa-a675512ec8f1-kube-api-access-nhdcr\") pod \"service-ca-operator-777779d784-wn86t\" (UID: \"991a9129-37d1-4407-a9aa-a675512ec8f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn86t" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.257599 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p6gf\" (UniqueName: \"kubernetes.io/projected/571d990f-04d9-4e62-902b-f44b7b666901-kube-api-access-6p6gf\") pod \"olm-operator-6b444d44fb-k2lwz\" (UID: \"571d990f-04d9-4e62-902b-f44b7b666901\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k2lwz" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.257749 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6e5e0a0-6ff8-44eb-99f6-eb9a4eb2aa47-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8x7bf\" (UID: \"e6e5e0a0-6ff8-44eb-99f6-eb9a4eb2aa47\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8x7bf" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.257796 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw8rq\" (UniqueName: \"kubernetes.io/projected/aa9441a2-a8c5-4f10-89c6-88fa689e9f7c-kube-api-access-vw8rq\") pod \"openshift-config-operator-7777fb866f-4lflr\" (UID: \"aa9441a2-a8c5-4f10-89c6-88fa689e9f7c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4lflr" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.257833 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9c926cbb-e6ad-4e49-9a6c-afaf6eb9ced6-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-ljhpv\" (UID: \"9c926cbb-e6ad-4e49-9a6c-afaf6eb9ced6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ljhpv" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.257856 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-84w94\" (UID: \"30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84w94" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.257879 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/51d519c2-0060-4aa2-9b25-11c742aca2d1-signing-cabundle\") pod \"service-ca-9c57cc56f-9wxlv\" (UID: \"51d519c2-0060-4aa2-9b25-11c742aca2d1\") " pod="openshift-service-ca/service-ca-9c57cc56f-9wxlv" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.258771 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/70fe2241-a755-4556-a84a-733306f90d66-trusted-ca\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.260956 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6e73973b-70de-449a-9ad2-b965b7a83e1f-certs\") pod \"machine-config-server-wmjcx\" (UID: \"6e73973b-70de-449a-9ad2-b965b7a83e1f\") " pod="openshift-machine-config-operator/machine-config-server-wmjcx" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.261000 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/763388db-ca9e-432e-9c91-4821392af1a1-csi-data-dir\") pod \"csi-hostpathplugin-96w58\" (UID: \"763388db-ca9e-432e-9c91-4821392af1a1\") " pod="hostpath-provisioner/csi-hostpathplugin-96w58" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.261036 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/916530ba-091f-4181-aee2-eaa5cd546db5-config\") pod \"machine-approver-56656f9798-9c4d9\" (UID: \"916530ba-091f-4181-aee2-eaa5cd546db5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c4d9" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.261055 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c92jj\" (UniqueName: \"kubernetes.io/projected/763388db-ca9e-432e-9c91-4821392af1a1-kube-api-access-c92jj\") pod \"csi-hostpathplugin-96w58\" (UID: \"763388db-ca9e-432e-9c91-4821392af1a1\") " pod="hostpath-provisioner/csi-hostpathplugin-96w58" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.261105 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmx29\" (UniqueName: \"kubernetes.io/projected/e6e5e0a0-6ff8-44eb-99f6-eb9a4eb2aa47-kube-api-access-fmx29\") pod \"kube-storage-version-migrator-operator-b67b599dd-8x7bf\" (UID: \"e6e5e0a0-6ff8-44eb-99f6-eb9a4eb2aa47\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8x7bf" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.261173 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c1eea5fd-9b4d-4d4d-a715-6523309cdd87-profile-collector-cert\") pod \"catalog-operator-68c6474976-vj8ww\" (UID: \"c1eea5fd-9b4d-4d4d-a715-6523309cdd87\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vj8ww" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.261195 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/bd767915-7aeb-45a4-96a3-c2f636c5e754-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b7nbg\" (UID: \"bd767915-7aeb-45a4-96a3-c2f636c5e754\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7nbg" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.261226 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/70fe2241-a755-4556-a84a-733306f90d66-registry-certificates\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.261276 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gcll\" (UniqueName: \"kubernetes.io/projected/926501e1-d68a-4c86-b59e-f87c43c3a9ae-kube-api-access-4gcll\") pod \"packageserver-d55dfcdfc-vh7kj\" (UID: \"926501e1-d68a-4c86-b59e-f87c43c3a9ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.261316 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ef87f11-a2ee-48e5-86ec-94283d235933-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2bwvx\" (UID: \"4ef87f11-a2ee-48e5-86ec-94283d235933\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2bwvx" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.261344 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/571d990f-04d9-4e62-902b-f44b7b666901-profile-collector-cert\") pod \"olm-operator-6b444d44fb-k2lwz\" (UID: \"571d990f-04d9-4e62-902b-f44b7b666901\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k2lwz" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.261365 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9ab91fd8-27d8-458a-97b9-09b9ba008d67-cert\") pod \"ingress-canary-ps67b\" (UID: \"9ab91fd8-27d8-458a-97b9-09b9ba008d67\") " pod="openshift-ingress-canary/ingress-canary-ps67b" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.270218 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-84w94\" (UID: \"30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84w94" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.271250 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/acb21b38-5eb2-4db3-a8ba-9c022c9ae79e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-n8hmz\" (UID: \"acb21b38-5eb2-4db3-a8ba-9c022c9ae79e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n8hmz" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.271433 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-84w94\" (UID: \"30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84w94" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.271549 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/763388db-ca9e-432e-9c91-4821392af1a1-registration-dir\") pod \"csi-hostpathplugin-96w58\" (UID: \"763388db-ca9e-432e-9c91-4821392af1a1\") " pod="hostpath-provisioner/csi-hostpathplugin-96w58" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.273741 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/70fe2241-a755-4556-a84a-733306f90d66-registry-tls\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.274807 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/70fe2241-a755-4556-a84a-733306f90d66-registry-certificates\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.286519 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-84w94\" (UID: \"30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84w94" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.286859 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6e5e0a0-6ff8-44eb-99f6-eb9a4eb2aa47-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8x7bf\" (UID: \"e6e5e0a0-6ff8-44eb-99f6-eb9a4eb2aa47\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8x7bf" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.287095 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/70fe2241-a755-4556-a84a-733306f90d66-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.287259 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa9441a2-a8c5-4f10-89c6-88fa689e9f7c-serving-cert\") pod \"openshift-config-operator-7777fb866f-4lflr\" (UID: \"aa9441a2-a8c5-4f10-89c6-88fa689e9f7c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4lflr" Oct 13 08:46:56 crc kubenswrapper[4685]: W1013 08:46:56.292328 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44a7f769_9a31_4555_be16_51e310ac34e3.slice/crio-6f09dddfdd8e870a0bde510c22bd0eda40b185eed529e87e33ba20b1c43e7b31 WatchSource:0}: Error finding container 6f09dddfdd8e870a0bde510c22bd0eda40b185eed529e87e33ba20b1c43e7b31: Status 404 returned error can't find the container with id 6f09dddfdd8e870a0bde510c22bd0eda40b185eed529e87e33ba20b1c43e7b31 Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.303370 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-84w94\" (UID: \"30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84w94" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.304518 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvqmj\" (UniqueName: \"kubernetes.io/projected/30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90-kube-api-access-jvqmj\") pod \"cluster-image-registry-operator-dc59b4c8b-84w94\" (UID: \"30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84w94" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.322254 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/70fe2241-a755-4556-a84a-733306f90d66-bound-sa-token\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.335476 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kz5ht\" (UniqueName: \"kubernetes.io/projected/70fe2241-a755-4556-a84a-733306f90d66-kube-api-access-kz5ht\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.353573 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw8rq\" (UniqueName: \"kubernetes.io/projected/aa9441a2-a8c5-4f10-89c6-88fa689e9f7c-kube-api-access-vw8rq\") pod \"openshift-config-operator-7777fb866f-4lflr\" (UID: \"aa9441a2-a8c5-4f10-89c6-88fa689e9f7c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4lflr" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.358209 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-r96rq" event={"ID":"0fe547bf-ea91-4546-a886-c613abeae02a","Type":"ContainerStarted","Data":"f65a4507bc119d147d83f5add56a3c2e8c0eaa0ffdb2cdb69a4aec8117465612"} Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.360851 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rx99b" event={"ID":"b9f835b0-16cf-4f05-a8a2-2767a929ca25","Type":"ContainerStarted","Data":"9bb961f0bd899dc70d27b0664d8a529a52f27b54238ed0c048a76cf4263449d3"} Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.361655 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" event={"ID":"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d","Type":"ContainerStarted","Data":"9ea816bfb6ce025dde15f7c50f84ea4d100ed3ebecb13bc4b7802ee167833b1b"} Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.367808 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-2kbsj" event={"ID":"44a7f769-9a31-4555-be16-51e310ac34e3","Type":"ContainerStarted","Data":"6f09dddfdd8e870a0bde510c22bd0eda40b185eed529e87e33ba20b1c43e7b31"} Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.372458 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.372650 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gcll\" (UniqueName: \"kubernetes.io/projected/926501e1-d68a-4c86-b59e-f87c43c3a9ae-kube-api-access-4gcll\") pod \"packageserver-d55dfcdfc-vh7kj\" (UID: \"926501e1-d68a-4c86-b59e-f87c43c3a9ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.372680 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ef87f11-a2ee-48e5-86ec-94283d235933-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2bwvx\" (UID: \"4ef87f11-a2ee-48e5-86ec-94283d235933\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2bwvx" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.372706 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/571d990f-04d9-4e62-902b-f44b7b666901-profile-collector-cert\") pod \"olm-operator-6b444d44fb-k2lwz\" (UID: \"571d990f-04d9-4e62-902b-f44b7b666901\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k2lwz" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.372723 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9ab91fd8-27d8-458a-97b9-09b9ba008d67-cert\") pod \"ingress-canary-ps67b\" (UID: \"9ab91fd8-27d8-458a-97b9-09b9ba008d67\") " pod="openshift-ingress-canary/ingress-canary-ps67b" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.372751 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/acb21b38-5eb2-4db3-a8ba-9c022c9ae79e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-n8hmz\" (UID: \"acb21b38-5eb2-4db3-a8ba-9c022c9ae79e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n8hmz" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.372775 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/763388db-ca9e-432e-9c91-4821392af1a1-registration-dir\") pod \"csi-hostpathplugin-96w58\" (UID: \"763388db-ca9e-432e-9c91-4821392af1a1\") " pod="hostpath-provisioner/csi-hostpathplugin-96w58" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.372792 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8p545\" (UniqueName: \"kubernetes.io/projected/c1eea5fd-9b4d-4d4d-a715-6523309cdd87-kube-api-access-8p545\") pod \"catalog-operator-68c6474976-vj8ww\" (UID: \"c1eea5fd-9b4d-4d4d-a715-6523309cdd87\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vj8ww" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.372807 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/926501e1-d68a-4c86-b59e-f87c43c3a9ae-webhook-cert\") pod \"packageserver-d55dfcdfc-vh7kj\" (UID: \"926501e1-d68a-4c86-b59e-f87c43c3a9ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.372830 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ef87f11-a2ee-48e5-86ec-94283d235933-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2bwvx\" (UID: \"4ef87f11-a2ee-48e5-86ec-94283d235933\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2bwvx" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.372847 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/d869fa03-5196-4a23-a7d3-9bb709891678-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-4zvdw\" (UID: \"d869fa03-5196-4a23-a7d3-9bb709891678\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4zvdw" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.372865 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/53a5a736-ebd0-4c09-90b3-d70a4943c038-config-volume\") pod \"dns-default-cnh25\" (UID: \"53a5a736-ebd0-4c09-90b3-d70a4943c038\") " pod="openshift-dns/dns-default-cnh25" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.372879 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/571d990f-04d9-4e62-902b-f44b7b666901-srv-cert\") pod \"olm-operator-6b444d44fb-k2lwz\" (UID: \"571d990f-04d9-4e62-902b-f44b7b666901\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k2lwz" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.372898 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/763388db-ca9e-432e-9c91-4821392af1a1-plugins-dir\") pod \"csi-hostpathplugin-96w58\" (UID: \"763388db-ca9e-432e-9c91-4821392af1a1\") " pod="hostpath-provisioner/csi-hostpathplugin-96w58" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.372939 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c926cbb-e6ad-4e49-9a6c-afaf6eb9ced6-config\") pod \"kube-apiserver-operator-766d6c64bb-ljhpv\" (UID: \"9c926cbb-e6ad-4e49-9a6c-afaf6eb9ced6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ljhpv" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.372956 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b8ae1ce1-bf5f-4d13-87b4-0810f63434af-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7sx5f\" (UID: \"b8ae1ce1-bf5f-4d13-87b4-0810f63434af\") " pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.372980 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/763388db-ca9e-432e-9c91-4821392af1a1-socket-dir\") pod \"csi-hostpathplugin-96w58\" (UID: \"763388db-ca9e-432e-9c91-4821392af1a1\") " pod="hostpath-provisioner/csi-hostpathplugin-96w58" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.372996 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/916530ba-091f-4181-aee2-eaa5cd546db5-machine-approver-tls\") pod \"machine-approver-56656f9798-9c4d9\" (UID: \"916530ba-091f-4181-aee2-eaa5cd546db5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c4d9" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373015 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6vrw\" (UniqueName: \"kubernetes.io/projected/6e73973b-70de-449a-9ad2-b965b7a83e1f-kube-api-access-n6vrw\") pod \"machine-config-server-wmjcx\" (UID: \"6e73973b-70de-449a-9ad2-b965b7a83e1f\") " pod="openshift-machine-config-operator/machine-config-server-wmjcx" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.372990 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" event={"ID":"5072c69e-c292-478d-9998-8b5237dfc33f","Type":"ContainerStarted","Data":"151f27404ab90dc5db9ba8ef80f161f8af4bca5ce732becc722a3fc05e7504c4"} Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373059 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrpk2\" (UniqueName: \"kubernetes.io/projected/3672a087-9d9b-45ee-8f80-686959b395d6-kube-api-access-xrpk2\") pod \"collect-profiles-29339085-84tsn\" (UID: \"3672a087-9d9b-45ee-8f80-686959b395d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373076 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zbqc\" (UniqueName: \"kubernetes.io/projected/b8ae1ce1-bf5f-4d13-87b4-0810f63434af-kube-api-access-9zbqc\") pod \"marketplace-operator-79b997595-7sx5f\" (UID: \"b8ae1ce1-bf5f-4d13-87b4-0810f63434af\") " pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373093 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6e73973b-70de-449a-9ad2-b965b7a83e1f-node-bootstrap-token\") pod \"machine-config-server-wmjcx\" (UID: \"6e73973b-70de-449a-9ad2-b965b7a83e1f\") " pod="openshift-machine-config-operator/machine-config-server-wmjcx" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373106 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/53a5a736-ebd0-4c09-90b3-d70a4943c038-metrics-tls\") pod \"dns-default-cnh25\" (UID: \"53a5a736-ebd0-4c09-90b3-d70a4943c038\") " pod="openshift-dns/dns-default-cnh25" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373123 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/926501e1-d68a-4c86-b59e-f87c43c3a9ae-tmpfs\") pod \"packageserver-d55dfcdfc-vh7kj\" (UID: \"926501e1-d68a-4c86-b59e-f87c43c3a9ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373151 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztzr7\" (UniqueName: \"kubernetes.io/projected/51d519c2-0060-4aa2-9b25-11c742aca2d1-kube-api-access-ztzr7\") pod \"service-ca-9c57cc56f-9wxlv\" (UID: \"51d519c2-0060-4aa2-9b25-11c742aca2d1\") " pod="openshift-service-ca/service-ca-9c57cc56f-9wxlv" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373170 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/acb21b38-5eb2-4db3-a8ba-9c022c9ae79e-proxy-tls\") pod \"machine-config-controller-84d6567774-n8hmz\" (UID: \"acb21b38-5eb2-4db3-a8ba-9c022c9ae79e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n8hmz" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373184 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b8ae1ce1-bf5f-4d13-87b4-0810f63434af-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7sx5f\" (UID: \"b8ae1ce1-bf5f-4d13-87b4-0810f63434af\") " pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373203 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/763388db-ca9e-432e-9c91-4821392af1a1-mountpoint-dir\") pod \"csi-hostpathplugin-96w58\" (UID: \"763388db-ca9e-432e-9c91-4821392af1a1\") " pod="hostpath-provisioner/csi-hostpathplugin-96w58" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373223 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/926501e1-d68a-4c86-b59e-f87c43c3a9ae-apiservice-cert\") pod \"packageserver-d55dfcdfc-vh7kj\" (UID: \"926501e1-d68a-4c86-b59e-f87c43c3a9ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373241 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7kgm\" (UniqueName: \"kubernetes.io/projected/53a5a736-ebd0-4c09-90b3-d70a4943c038-kube-api-access-v7kgm\") pod \"dns-default-cnh25\" (UID: \"53a5a736-ebd0-4c09-90b3-d70a4943c038\") " pod="openshift-dns/dns-default-cnh25" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373258 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgtzx\" (UniqueName: \"kubernetes.io/projected/acb21b38-5eb2-4db3-a8ba-9c022c9ae79e-kube-api-access-jgtzx\") pod \"machine-config-controller-84d6567774-n8hmz\" (UID: \"acb21b38-5eb2-4db3-a8ba-9c022c9ae79e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n8hmz" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373281 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnldg\" (UniqueName: \"kubernetes.io/projected/47f6f699-d6e2-4289-bd8a-9ec387887523-kube-api-access-tnldg\") pod \"migrator-59844c95c7-xd6tf\" (UID: \"47f6f699-d6e2-4289-bd8a-9ec387887523\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xd6tf" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373299 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/916530ba-091f-4181-aee2-eaa5cd546db5-auth-proxy-config\") pod \"machine-approver-56656f9798-9c4d9\" (UID: \"916530ba-091f-4181-aee2-eaa5cd546db5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c4d9" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373316 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c926cbb-e6ad-4e49-9a6c-afaf6eb9ced6-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-ljhpv\" (UID: \"9c926cbb-e6ad-4e49-9a6c-afaf6eb9ced6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ljhpv" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373335 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ef87f11-a2ee-48e5-86ec-94283d235933-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2bwvx\" (UID: \"4ef87f11-a2ee-48e5-86ec-94283d235933\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2bwvx" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373357 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9cdz\" (UniqueName: \"kubernetes.io/projected/bd767915-7aeb-45a4-96a3-c2f636c5e754-kube-api-access-l9cdz\") pod \"package-server-manager-789f6589d5-b7nbg\" (UID: \"bd767915-7aeb-45a4-96a3-c2f636c5e754\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7nbg" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373376 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl4cc\" (UniqueName: \"kubernetes.io/projected/d869fa03-5196-4a23-a7d3-9bb709891678-kube-api-access-xl4cc\") pod \"control-plane-machine-set-operator-78cbb6b69f-4zvdw\" (UID: \"d869fa03-5196-4a23-a7d3-9bb709891678\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4zvdw" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373393 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9cjk\" (UniqueName: \"kubernetes.io/projected/9ab91fd8-27d8-458a-97b9-09b9ba008d67-kube-api-access-p9cjk\") pod \"ingress-canary-ps67b\" (UID: \"9ab91fd8-27d8-458a-97b9-09b9ba008d67\") " pod="openshift-ingress-canary/ingress-canary-ps67b" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373409 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c1eea5fd-9b4d-4d4d-a715-6523309cdd87-srv-cert\") pod \"catalog-operator-68c6474976-vj8ww\" (UID: \"c1eea5fd-9b4d-4d4d-a715-6523309cdd87\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vj8ww" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373428 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/991a9129-37d1-4407-a9aa-a675512ec8f1-config\") pod \"service-ca-operator-777779d784-wn86t\" (UID: \"991a9129-37d1-4407-a9aa-a675512ec8f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn86t" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373447 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3672a087-9d9b-45ee-8f80-686959b395d6-config-volume\") pod \"collect-profiles-29339085-84tsn\" (UID: \"3672a087-9d9b-45ee-8f80-686959b395d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373468 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq2tm\" (UniqueName: \"kubernetes.io/projected/916530ba-091f-4181-aee2-eaa5cd546db5-kube-api-access-xq2tm\") pod \"machine-approver-56656f9798-9c4d9\" (UID: \"916530ba-091f-4181-aee2-eaa5cd546db5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c4d9" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373485 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3672a087-9d9b-45ee-8f80-686959b395d6-secret-volume\") pod \"collect-profiles-29339085-84tsn\" (UID: \"3672a087-9d9b-45ee-8f80-686959b395d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373503 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/991a9129-37d1-4407-a9aa-a675512ec8f1-serving-cert\") pod \"service-ca-operator-777779d784-wn86t\" (UID: \"991a9129-37d1-4407-a9aa-a675512ec8f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn86t" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373518 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/51d519c2-0060-4aa2-9b25-11c742aca2d1-signing-key\") pod \"service-ca-9c57cc56f-9wxlv\" (UID: \"51d519c2-0060-4aa2-9b25-11c742aca2d1\") " pod="openshift-service-ca/service-ca-9c57cc56f-9wxlv" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373535 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhdcr\" (UniqueName: \"kubernetes.io/projected/991a9129-37d1-4407-a9aa-a675512ec8f1-kube-api-access-nhdcr\") pod \"service-ca-operator-777779d784-wn86t\" (UID: \"991a9129-37d1-4407-a9aa-a675512ec8f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn86t" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373553 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p6gf\" (UniqueName: \"kubernetes.io/projected/571d990f-04d9-4e62-902b-f44b7b666901-kube-api-access-6p6gf\") pod \"olm-operator-6b444d44fb-k2lwz\" (UID: \"571d990f-04d9-4e62-902b-f44b7b666901\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k2lwz" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373586 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9c926cbb-e6ad-4e49-9a6c-afaf6eb9ced6-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-ljhpv\" (UID: \"9c926cbb-e6ad-4e49-9a6c-afaf6eb9ced6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ljhpv" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373605 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/51d519c2-0060-4aa2-9b25-11c742aca2d1-signing-cabundle\") pod \"service-ca-9c57cc56f-9wxlv\" (UID: \"51d519c2-0060-4aa2-9b25-11c742aca2d1\") " pod="openshift-service-ca/service-ca-9c57cc56f-9wxlv" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373623 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6e73973b-70de-449a-9ad2-b965b7a83e1f-certs\") pod \"machine-config-server-wmjcx\" (UID: \"6e73973b-70de-449a-9ad2-b965b7a83e1f\") " pod="openshift-machine-config-operator/machine-config-server-wmjcx" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373637 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/763388db-ca9e-432e-9c91-4821392af1a1-csi-data-dir\") pod \"csi-hostpathplugin-96w58\" (UID: \"763388db-ca9e-432e-9c91-4821392af1a1\") " pod="hostpath-provisioner/csi-hostpathplugin-96w58" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373652 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/916530ba-091f-4181-aee2-eaa5cd546db5-config\") pod \"machine-approver-56656f9798-9c4d9\" (UID: \"916530ba-091f-4181-aee2-eaa5cd546db5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c4d9" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.373668 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c92jj\" (UniqueName: \"kubernetes.io/projected/763388db-ca9e-432e-9c91-4821392af1a1-kube-api-access-c92jj\") pod \"csi-hostpathplugin-96w58\" (UID: \"763388db-ca9e-432e-9c91-4821392af1a1\") " pod="hostpath-provisioner/csi-hostpathplugin-96w58" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.374438 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:56 crc kubenswrapper[4685]: E1013 08:46:56.375905 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:46:56.875872691 +0000 UTC m=+142.023748452 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.376357 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c1eea5fd-9b4d-4d4d-a715-6523309cdd87-profile-collector-cert\") pod \"catalog-operator-68c6474976-vj8ww\" (UID: \"c1eea5fd-9b4d-4d4d-a715-6523309cdd87\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vj8ww" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.376455 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/bd767915-7aeb-45a4-96a3-c2f636c5e754-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b7nbg\" (UID: \"bd767915-7aeb-45a4-96a3-c2f636c5e754\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7nbg" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.377006 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/763388db-ca9e-432e-9c91-4821392af1a1-plugins-dir\") pod \"csi-hostpathplugin-96w58\" (UID: \"763388db-ca9e-432e-9c91-4821392af1a1\") " pod="hostpath-provisioner/csi-hostpathplugin-96w58" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.377524 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/763388db-ca9e-432e-9c91-4821392af1a1-registration-dir\") pod \"csi-hostpathplugin-96w58\" (UID: \"763388db-ca9e-432e-9c91-4821392af1a1\") " pod="hostpath-provisioner/csi-hostpathplugin-96w58" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.377829 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/916530ba-091f-4181-aee2-eaa5cd546db5-auth-proxy-config\") pod \"machine-approver-56656f9798-9c4d9\" (UID: \"916530ba-091f-4181-aee2-eaa5cd546db5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c4d9" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.377840 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/763388db-ca9e-432e-9c91-4821392af1a1-socket-dir\") pod \"csi-hostpathplugin-96w58\" (UID: \"763388db-ca9e-432e-9c91-4821392af1a1\") " pod="hostpath-provisioner/csi-hostpathplugin-96w58" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.379760 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b8ae1ce1-bf5f-4d13-87b4-0810f63434af-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7sx5f\" (UID: \"b8ae1ce1-bf5f-4d13-87b4-0810f63434af\") " pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.384461 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/763388db-ca9e-432e-9c91-4821392af1a1-mountpoint-dir\") pod \"csi-hostpathplugin-96w58\" (UID: \"763388db-ca9e-432e-9c91-4821392af1a1\") " pod="hostpath-provisioner/csi-hostpathplugin-96w58" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.386052 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/53a5a736-ebd0-4c09-90b3-d70a4943c038-config-volume\") pod \"dns-default-cnh25\" (UID: \"53a5a736-ebd0-4c09-90b3-d70a4943c038\") " pod="openshift-dns/dns-default-cnh25" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.388648 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/acb21b38-5eb2-4db3-a8ba-9c022c9ae79e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-n8hmz\" (UID: \"acb21b38-5eb2-4db3-a8ba-9c022c9ae79e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n8hmz" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.389281 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/763388db-ca9e-432e-9c91-4821392af1a1-csi-data-dir\") pod \"csi-hostpathplugin-96w58\" (UID: \"763388db-ca9e-432e-9c91-4821392af1a1\") " pod="hostpath-provisioner/csi-hostpathplugin-96w58" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.389512 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ef87f11-a2ee-48e5-86ec-94283d235933-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2bwvx\" (UID: \"4ef87f11-a2ee-48e5-86ec-94283d235933\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2bwvx" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.393387 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmx29\" (UniqueName: \"kubernetes.io/projected/e6e5e0a0-6ff8-44eb-99f6-eb9a4eb2aa47-kube-api-access-fmx29\") pod \"kube-storage-version-migrator-operator-b67b599dd-8x7bf\" (UID: \"e6e5e0a0-6ff8-44eb-99f6-eb9a4eb2aa47\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8x7bf" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.394287 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/926501e1-d68a-4c86-b59e-f87c43c3a9ae-tmpfs\") pod \"packageserver-d55dfcdfc-vh7kj\" (UID: \"926501e1-d68a-4c86-b59e-f87c43c3a9ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.401496 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/916530ba-091f-4181-aee2-eaa5cd546db5-config\") pod \"machine-approver-56656f9798-9c4d9\" (UID: \"916530ba-091f-4181-aee2-eaa5cd546db5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c4d9" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.402556 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/51d519c2-0060-4aa2-9b25-11c742aca2d1-signing-cabundle\") pod \"service-ca-9c57cc56f-9wxlv\" (UID: \"51d519c2-0060-4aa2-9b25-11c742aca2d1\") " pod="openshift-service-ca/service-ca-9c57cc56f-9wxlv" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.403380 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/571d990f-04d9-4e62-902b-f44b7b666901-profile-collector-cert\") pod \"olm-operator-6b444d44fb-k2lwz\" (UID: \"571d990f-04d9-4e62-902b-f44b7b666901\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k2lwz" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.403974 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/571d990f-04d9-4e62-902b-f44b7b666901-srv-cert\") pod \"olm-operator-6b444d44fb-k2lwz\" (UID: \"571d990f-04d9-4e62-902b-f44b7b666901\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k2lwz" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.404468 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9ab91fd8-27d8-458a-97b9-09b9ba008d67-cert\") pod \"ingress-canary-ps67b\" (UID: \"9ab91fd8-27d8-458a-97b9-09b9ba008d67\") " pod="openshift-ingress-canary/ingress-canary-ps67b" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.404666 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/926501e1-d68a-4c86-b59e-f87c43c3a9ae-apiservice-cert\") pod \"packageserver-d55dfcdfc-vh7kj\" (UID: \"926501e1-d68a-4c86-b59e-f87c43c3a9ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.405280 4685 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-54dh8 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.6:6443/healthz\": dial tcp 10.217.0.6:6443: connect: connection refused" start-of-body= Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.405410 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" podUID="5072c69e-c292-478d-9998-8b5237dfc33f" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.6:6443/healthz\": dial tcp 10.217.0.6:6443: connect: connection refused" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.407545 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/acb21b38-5eb2-4db3-a8ba-9c022c9ae79e-proxy-tls\") pod \"machine-config-controller-84d6567774-n8hmz\" (UID: \"acb21b38-5eb2-4db3-a8ba-9c022c9ae79e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n8hmz" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.408643 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/bd767915-7aeb-45a4-96a3-c2f636c5e754-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b7nbg\" (UID: \"bd767915-7aeb-45a4-96a3-c2f636c5e754\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7nbg" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.409228 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/916530ba-091f-4181-aee2-eaa5cd546db5-machine-approver-tls\") pod \"machine-approver-56656f9798-9c4d9\" (UID: \"916530ba-091f-4181-aee2-eaa5cd546db5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c4d9" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.410119 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ef87f11-a2ee-48e5-86ec-94283d235933-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2bwvx\" (UID: \"4ef87f11-a2ee-48e5-86ec-94283d235933\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2bwvx" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.411122 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qm75h" event={"ID":"76cb8882-50e3-4443-b73e-27dbb83e4173","Type":"ContainerStarted","Data":"4a71ab0974c29ffb4163fa52928068ea3799086e7460579e8ddbe9f3f5db3447"} Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.412477 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/991a9129-37d1-4407-a9aa-a675512ec8f1-config\") pod \"service-ca-operator-777779d784-wn86t\" (UID: \"991a9129-37d1-4407-a9aa-a675512ec8f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn86t" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.413049 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/d869fa03-5196-4a23-a7d3-9bb709891678-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-4zvdw\" (UID: \"d869fa03-5196-4a23-a7d3-9bb709891678\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4zvdw" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.413818 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/53a5a736-ebd0-4c09-90b3-d70a4943c038-metrics-tls\") pod \"dns-default-cnh25\" (UID: \"53a5a736-ebd0-4c09-90b3-d70a4943c038\") " pod="openshift-dns/dns-default-cnh25" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.413851 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3672a087-9d9b-45ee-8f80-686959b395d6-config-volume\") pod \"collect-profiles-29339085-84tsn\" (UID: \"3672a087-9d9b-45ee-8f80-686959b395d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.414039 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c926cbb-e6ad-4e49-9a6c-afaf6eb9ced6-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-ljhpv\" (UID: \"9c926cbb-e6ad-4e49-9a6c-afaf6eb9ced6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ljhpv" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.414224 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c1eea5fd-9b4d-4d4d-a715-6523309cdd87-srv-cert\") pod \"catalog-operator-68c6474976-vj8ww\" (UID: \"c1eea5fd-9b4d-4d4d-a715-6523309cdd87\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vj8ww" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.414546 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/51d519c2-0060-4aa2-9b25-11c742aca2d1-signing-key\") pod \"service-ca-9c57cc56f-9wxlv\" (UID: \"51d519c2-0060-4aa2-9b25-11c742aca2d1\") " pod="openshift-service-ca/service-ca-9c57cc56f-9wxlv" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.415313 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84w94" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.419001 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6e73973b-70de-449a-9ad2-b965b7a83e1f-certs\") pod \"machine-config-server-wmjcx\" (UID: \"6e73973b-70de-449a-9ad2-b965b7a83e1f\") " pod="openshift-machine-config-operator/machine-config-server-wmjcx" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.421025 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6e73973b-70de-449a-9ad2-b965b7a83e1f-node-bootstrap-token\") pod \"machine-config-server-wmjcx\" (UID: \"6e73973b-70de-449a-9ad2-b965b7a83e1f\") " pod="openshift-machine-config-operator/machine-config-server-wmjcx" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.421411 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c926cbb-e6ad-4e49-9a6c-afaf6eb9ced6-config\") pod \"kube-apiserver-operator-766d6c64bb-ljhpv\" (UID: \"9c926cbb-e6ad-4e49-9a6c-afaf6eb9ced6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ljhpv" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.427157 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b8ae1ce1-bf5f-4d13-87b4-0810f63434af-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7sx5f\" (UID: \"b8ae1ce1-bf5f-4d13-87b4-0810f63434af\") " pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.417195 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/926501e1-d68a-4c86-b59e-f87c43c3a9ae-webhook-cert\") pod \"packageserver-d55dfcdfc-vh7kj\" (UID: \"926501e1-d68a-4c86-b59e-f87c43c3a9ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.428670 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3672a087-9d9b-45ee-8f80-686959b395d6-secret-volume\") pod \"collect-profiles-29339085-84tsn\" (UID: \"3672a087-9d9b-45ee-8f80-686959b395d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.432659 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c1eea5fd-9b4d-4d4d-a715-6523309cdd87-profile-collector-cert\") pod \"catalog-operator-68c6474976-vj8ww\" (UID: \"c1eea5fd-9b4d-4d4d-a715-6523309cdd87\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vj8ww" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.433337 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrpk2\" (UniqueName: \"kubernetes.io/projected/3672a087-9d9b-45ee-8f80-686959b395d6-kube-api-access-xrpk2\") pod \"collect-profiles-29339085-84tsn\" (UID: \"3672a087-9d9b-45ee-8f80-686959b395d6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.435274 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4lflr" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.438653 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8p545\" (UniqueName: \"kubernetes.io/projected/c1eea5fd-9b4d-4d4d-a715-6523309cdd87-kube-api-access-8p545\") pod \"catalog-operator-68c6474976-vj8ww\" (UID: \"c1eea5fd-9b4d-4d4d-a715-6523309cdd87\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vj8ww" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.443466 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/991a9129-37d1-4407-a9aa-a675512ec8f1-serving-cert\") pod \"service-ca-operator-777779d784-wn86t\" (UID: \"991a9129-37d1-4407-a9aa-a675512ec8f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn86t" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.444480 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8x7bf" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.448524 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gpr4j" event={"ID":"887c5808-d4ac-4517-a1b9-1601f79aafa8","Type":"ContainerStarted","Data":"2468a5b52bcf91881a87774c457a0fda80fd2f09cfb273fc8d72cd3935768bd7"} Oct 13 08:46:56 crc kubenswrapper[4685]: W1013 08:46:56.456799 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda185671e_0a3a_4a9e_a884_65b448b1e922.slice/crio-8c0d36780f1d77ef9f103b28c727c0520beb6e327923f69f8ca9ec19678f4c45 WatchSource:0}: Error finding container 8c0d36780f1d77ef9f103b28c727c0520beb6e327923f69f8ca9ec19678f4c45: Status 404 returned error can't find the container with id 8c0d36780f1d77ef9f103b28c727c0520beb6e327923f69f8ca9ec19678f4c45 Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.462708 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7kgm\" (UniqueName: \"kubernetes.io/projected/53a5a736-ebd0-4c09-90b3-d70a4943c038-kube-api-access-v7kgm\") pod \"dns-default-cnh25\" (UID: \"53a5a736-ebd0-4c09-90b3-d70a4943c038\") " pod="openshift-dns/dns-default-cnh25" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.470827 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-gx6qj" event={"ID":"16484304-6976-45e8-a495-f1d2ee367f0f","Type":"ContainerStarted","Data":"eb84cf90b0898dd2fc99fab0c114053a787806daf85967a3410211428a74a350"} Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.470858 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-gx6qj" event={"ID":"16484304-6976-45e8-a495-f1d2ee367f0f","Type":"ContainerStarted","Data":"46405a1c5e6d6de272837e94e5d775db5fbb4d672663784fb24a92c82bb93021"} Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.479493 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: E1013 08:46:56.480136 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:46:56.980052115 +0000 UTC m=+142.127927956 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.490611 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgtzx\" (UniqueName: \"kubernetes.io/projected/acb21b38-5eb2-4db3-a8ba-9c022c9ae79e-kube-api-access-jgtzx\") pod \"machine-config-controller-84d6567774-n8hmz\" (UID: \"acb21b38-5eb2-4db3-a8ba-9c022c9ae79e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n8hmz" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.499363 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-9cnpf" event={"ID":"bcca1906-64bf-44dc-9307-b8d44d2b506f","Type":"ContainerStarted","Data":"090c325a47c8dd18895f58d9f6e93ed6bcc1d1731013b6d2fb1a472a2aa192dd"} Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.499822 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n8hmz" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.505793 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnldg\" (UniqueName: \"kubernetes.io/projected/47f6f699-d6e2-4289-bd8a-9ec387887523-kube-api-access-tnldg\") pod \"migrator-59844c95c7-xd6tf\" (UID: \"47f6f699-d6e2-4289-bd8a-9ec387887523\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xd6tf" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.515723 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vj8ww" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.522455 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gcll\" (UniqueName: \"kubernetes.io/projected/926501e1-d68a-4c86-b59e-f87c43c3a9ae-kube-api-access-4gcll\") pod \"packageserver-d55dfcdfc-vh7kj\" (UID: \"926501e1-d68a-4c86-b59e-f87c43c3a9ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.522714 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pbpvq" event={"ID":"c2f79e15-fea6-4391-a711-53ca225b1857","Type":"ContainerStarted","Data":"03f054ee1171699f17098ff111d201cee15343a67a8cb1cc072517531d3ed031"} Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.528091 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-lcljr"] Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.533876 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.541400 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ef87f11-a2ee-48e5-86ec-94283d235933-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2bwvx\" (UID: \"4ef87f11-a2ee-48e5-86ec-94283d235933\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2bwvx" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.556339 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" event={"ID":"b57dcc9c-6002-4c47-ae85-c287659cb679","Type":"ContainerStarted","Data":"4de31c7db49bd4a2134c1bb5fc7d484724396de62d0523094f46f6411ef74b79"} Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.561964 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zbqc\" (UniqueName: \"kubernetes.io/projected/b8ae1ce1-bf5f-4d13-87b4-0810f63434af-kube-api-access-9zbqc\") pod \"marketplace-operator-79b997595-7sx5f\" (UID: \"b8ae1ce1-bf5f-4d13-87b4-0810f63434af\") " pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.576977 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2bwvx" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.580947 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.581480 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztzr7\" (UniqueName: \"kubernetes.io/projected/51d519c2-0060-4aa2-9b25-11c742aca2d1-kube-api-access-ztzr7\") pod \"service-ca-9c57cc56f-9wxlv\" (UID: \"51d519c2-0060-4aa2-9b25-11c742aca2d1\") " pod="openshift-service-ca/service-ca-9c57cc56f-9wxlv" Oct 13 08:46:56 crc kubenswrapper[4685]: E1013 08:46:56.581959 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:46:57.081899491 +0000 UTC m=+142.229775252 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.584004 4685 generic.go:334] "Generic (PLEG): container finished" podID="8f247d85-48fa-49a1-9c0c-3ec7b7d202c0" containerID="fb6fa9ec79af02eb4cec001de0e7f865fa85d8de2fc0992db79f9b79c8fdb6d3" exitCode=0 Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.584362 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" event={"ID":"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0","Type":"ContainerDied","Data":"fb6fa9ec79af02eb4cec001de0e7f865fa85d8de2fc0992db79f9b79c8fdb6d3"} Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.589240 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-8ks9s" event={"ID":"83d9c4d8-10b5-4d1e-a71e-e6e92efd8379","Type":"ContainerStarted","Data":"9c1456cbc25366b53684cc9806bf2807288da71e3ee05ac1baa91848135160f1"} Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.605193 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" event={"ID":"7a81f8b4-7529-45f7-8ea5-df3252d70653","Type":"ContainerStarted","Data":"a66adfc40f59a9f4657f087d72d56f04b48cd8dce962d4a760458c5c54315485"} Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.605253 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" event={"ID":"7a81f8b4-7529-45f7-8ea5-df3252d70653","Type":"ContainerStarted","Data":"20b415acc33a00de655d765eec169b0519ed69c6923a8371aac4eb82f8cf502a"} Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.606223 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.616673 4685 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-9z47r container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.616717 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" podUID="7a81f8b4-7529-45f7-8ea5-df3252d70653" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.617163 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.623552 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhdcr\" (UniqueName: \"kubernetes.io/projected/991a9129-37d1-4407-a9aa-a675512ec8f1-kube-api-access-nhdcr\") pod \"service-ca-operator-777779d784-wn86t\" (UID: \"991a9129-37d1-4407-a9aa-a675512ec8f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn86t" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.624207 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6v84k" event={"ID":"3937dccc-95e5-4aa3-be81-6d12c31f7cb9","Type":"ContainerStarted","Data":"e89f2c762fd9ec1e86995b9cab85ceca608f8cda31427fffaf57aca12543d9b6"} Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.625851 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-cnh25" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.626319 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p6gf\" (UniqueName: \"kubernetes.io/projected/571d990f-04d9-4e62-902b-f44b7b666901-kube-api-access-6p6gf\") pod \"olm-operator-6b444d44fb-k2lwz\" (UID: \"571d990f-04d9-4e62-902b-f44b7b666901\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k2lwz" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.626405 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pk75f" event={"ID":"25a48963-29f5-4d84-91cf-b80e380ed9f1","Type":"ContainerStarted","Data":"a2c61f9d8ec4b584d1c156f4490eeed8364af04ca4e56606955e47e851c799ac"} Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.648291 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-csgb2" event={"ID":"e8d90978-a395-4313-bcb2-ca35a303b518","Type":"ContainerStarted","Data":"2a28ee85519838a0eaf4a4d4f0ae0cb5943a2f1974fa2a83eed5636b69a715bd"} Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.656206 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9c926cbb-e6ad-4e49-9a6c-afaf6eb9ced6-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-ljhpv\" (UID: \"9c926cbb-e6ad-4e49-9a6c-afaf6eb9ced6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ljhpv" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.661244 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6vrw\" (UniqueName: \"kubernetes.io/projected/6e73973b-70de-449a-9ad2-b965b7a83e1f-kube-api-access-n6vrw\") pod \"machine-config-server-wmjcx\" (UID: \"6e73973b-70de-449a-9ad2-b965b7a83e1f\") " pod="openshift-machine-config-operator/machine-config-server-wmjcx" Oct 13 08:46:56 crc kubenswrapper[4685]: W1013 08:46:56.667176 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc96407ea_0763_479d_8020_ace1607bd427.slice/crio-3a4949b90131c665e0a2c6bb16bd4664315c199a7e1989fc22998866609c01f3 WatchSource:0}: Error finding container 3a4949b90131c665e0a2c6bb16bd4664315c199a7e1989fc22998866609c01f3: Status 404 returned error can't find the container with id 3a4949b90131c665e0a2c6bb16bd4664315c199a7e1989fc22998866609c01f3 Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.682706 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: E1013 08:46:56.684938 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:46:57.1849071 +0000 UTC m=+142.332782861 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.736033 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9cdz\" (UniqueName: \"kubernetes.io/projected/bd767915-7aeb-45a4-96a3-c2f636c5e754-kube-api-access-l9cdz\") pod \"package-server-manager-789f6589d5-b7nbg\" (UID: \"bd767915-7aeb-45a4-96a3-c2f636c5e754\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7nbg" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.750410 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c92jj\" (UniqueName: \"kubernetes.io/projected/763388db-ca9e-432e-9c91-4821392af1a1-kube-api-access-c92jj\") pod \"csi-hostpathplugin-96w58\" (UID: \"763388db-ca9e-432e-9c91-4821392af1a1\") " pod="hostpath-provisioner/csi-hostpathplugin-96w58" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.754685 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl4cc\" (UniqueName: \"kubernetes.io/projected/d869fa03-5196-4a23-a7d3-9bb709891678-kube-api-access-xl4cc\") pod \"control-plane-machine-set-operator-78cbb6b69f-4zvdw\" (UID: \"d869fa03-5196-4a23-a7d3-9bb709891678\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4zvdw" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.770361 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9cjk\" (UniqueName: \"kubernetes.io/projected/9ab91fd8-27d8-458a-97b9-09b9ba008d67-kube-api-access-p9cjk\") pod \"ingress-canary-ps67b\" (UID: \"9ab91fd8-27d8-458a-97b9-09b9ba008d67\") " pod="openshift-ingress-canary/ingress-canary-ps67b" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.785543 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:46:56 crc kubenswrapper[4685]: E1013 08:46:56.785864 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:46:57.285847278 +0000 UTC m=+142.433723039 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.791735 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ljhpv" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.792037 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq2tm\" (UniqueName: \"kubernetes.io/projected/916530ba-091f-4181-aee2-eaa5cd546db5-kube-api-access-xq2tm\") pod \"machine-approver-56656f9798-9c4d9\" (UID: \"916530ba-091f-4181-aee2-eaa5cd546db5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c4d9" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.814633 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xd6tf" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.826439 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k2lwz" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.845199 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-4lflr"] Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.846051 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4zvdw" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.863501 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7nbg" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.863659 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.868461 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-9wxlv" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.888941 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:56 crc kubenswrapper[4685]: E1013 08:46:56.889309 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:46:57.389295391 +0000 UTC m=+142.537171152 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.918418 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn86t" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.918823 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-wmjcx" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.919473 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-96w58" Oct 13 08:46:56 crc kubenswrapper[4685]: I1013 08:46:56.937031 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-ps67b" Oct 13 08:46:57 crc kubenswrapper[4685]: W1013 08:46:56.990221 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa9441a2_a8c5_4f10_89c6_88fa689e9f7c.slice/crio-c553f1f843772d46f0f1dc032a7e7e48bc4a10b050b9fe64a58cbf3cd775adf2 WatchSource:0}: Error finding container c553f1f843772d46f0f1dc032a7e7e48bc4a10b050b9fe64a58cbf3cd775adf2: Status 404 returned error can't find the container with id c553f1f843772d46f0f1dc032a7e7e48bc4a10b050b9fe64a58cbf3cd775adf2 Oct 13 08:46:57 crc kubenswrapper[4685]: I1013 08:46:56.990836 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:46:57 crc kubenswrapper[4685]: E1013 08:46:56.994868 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:46:57.494853437 +0000 UTC m=+142.642729198 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:57 crc kubenswrapper[4685]: I1013 08:46:57.089538 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c4d9" Oct 13 08:46:57 crc kubenswrapper[4685]: I1013 08:46:57.095787 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:57 crc kubenswrapper[4685]: E1013 08:46:57.096137 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:46:57.596123915 +0000 UTC m=+142.743999676 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:57 crc kubenswrapper[4685]: I1013 08:46:57.207013 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:46:57 crc kubenswrapper[4685]: E1013 08:46:57.207806 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:46:57.707789081 +0000 UTC m=+142.855664842 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:57 crc kubenswrapper[4685]: I1013 08:46:57.311309 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:57 crc kubenswrapper[4685]: E1013 08:46:57.311656 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:46:57.811641575 +0000 UTC m=+142.959517326 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:57 crc kubenswrapper[4685]: I1013 08:46:57.416733 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:46:57 crc kubenswrapper[4685]: E1013 08:46:57.417396 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:46:57.917368606 +0000 UTC m=+143.065244367 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:57 crc kubenswrapper[4685]: I1013 08:46:57.518724 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:57 crc kubenswrapper[4685]: E1013 08:46:57.519586 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:46:58.019573491 +0000 UTC m=+143.167449252 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:57 crc kubenswrapper[4685]: I1013 08:46:57.622514 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84w94"] Oct 13 08:46:57 crc kubenswrapper[4685]: I1013 08:46:57.622659 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:46:57 crc kubenswrapper[4685]: E1013 08:46:57.623061 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:46:58.123037434 +0000 UTC m=+143.270913195 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:57 crc kubenswrapper[4685]: W1013 08:46:57.689226 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e73973b_70de_449a_9ad2_b965b7a83e1f.slice/crio-3118040bb76cd3523d2f71f2d6ed9b1c39b4831e763e49e95de403da4a374617 WatchSource:0}: Error finding container 3118040bb76cd3523d2f71f2d6ed9b1c39b4831e763e49e95de403da4a374617: Status 404 returned error can't find the container with id 3118040bb76cd3523d2f71f2d6ed9b1c39b4831e763e49e95de403da4a374617 Oct 13 08:46:57 crc kubenswrapper[4685]: I1013 08:46:57.723421 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-lcljr" event={"ID":"c96407ea-0763-479d-8020-ace1607bd427","Type":"ContainerStarted","Data":"3a4949b90131c665e0a2c6bb16bd4664315c199a7e1989fc22998866609c01f3"} Oct 13 08:46:57 crc kubenswrapper[4685]: I1013 08:46:57.725438 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:57 crc kubenswrapper[4685]: E1013 08:46:57.725839 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:46:58.225823158 +0000 UTC m=+143.373698919 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:57 crc kubenswrapper[4685]: I1013 08:46:57.759988 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gpr4j" event={"ID":"887c5808-d4ac-4517-a1b9-1601f79aafa8","Type":"ContainerStarted","Data":"791f3e6c9d2bc2f70bd972ca0276325f3dc2910e2f9ec0fc0f8468544f8d0070"} Oct 13 08:46:57 crc kubenswrapper[4685]: I1013 08:46:57.774091 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2bwvx"] Oct 13 08:46:57 crc kubenswrapper[4685]: I1013 08:46:57.797179 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8x7bf"] Oct 13 08:46:57 crc kubenswrapper[4685]: I1013 08:46:57.856609 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:46:57 crc kubenswrapper[4685]: E1013 08:46:57.859474 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:46:58.359449274 +0000 UTC m=+143.507325035 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:57 crc kubenswrapper[4685]: I1013 08:46:57.948192 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pbpvq" event={"ID":"c2f79e15-fea6-4391-a711-53ca225b1857","Type":"ContainerStarted","Data":"80358432542b391d5b285da8965cbadfebdd1d5ee264a0fbfde8db7e6ebeb98e"} Oct 13 08:46:57 crc kubenswrapper[4685]: I1013 08:46:57.958894 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:57 crc kubenswrapper[4685]: E1013 08:46:57.959287 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:46:58.459272459 +0000 UTC m=+143.607148220 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.028017 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn"] Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.060461 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:46:58 crc kubenswrapper[4685]: E1013 08:46:58.060648 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:46:58.560620759 +0000 UTC m=+143.708496520 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.061193 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:58 crc kubenswrapper[4685]: E1013 08:46:58.061742 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:46:58.561715132 +0000 UTC m=+143.709590893 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.123825 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qm75h" podStartSLOduration=122.12380809 podStartE2EDuration="2m2.12380809s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:58.105175558 +0000 UTC m=+143.253051339" watchObservedRunningTime="2025-10-13 08:46:58.12380809 +0000 UTC m=+143.271683851" Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.170473 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:46:58 crc kubenswrapper[4685]: E1013 08:46:58.171152 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:46:58.671137252 +0000 UTC m=+143.819013013 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.186116 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j6kt6" event={"ID":"a185671e-0a3a-4a9e-a884-65b448b1e922","Type":"ContainerStarted","Data":"2e761e980ae0b912f53cd6eb611d58d86dda29c1b024bfce54d699fb3ce688a2"} Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.186176 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j6kt6" event={"ID":"a185671e-0a3a-4a9e-a884-65b448b1e922","Type":"ContainerStarted","Data":"8c0d36780f1d77ef9f103b28c727c0520beb6e327923f69f8ca9ec19678f4c45"} Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.209888 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pk75f" event={"ID":"25a48963-29f5-4d84-91cf-b80e380ed9f1","Type":"ContainerStarted","Data":"d8556b5222456e9da29121bf78ec8192c91cc72285e8a3f36776fec0c4a3fb1d"} Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.228082 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4lflr" event={"ID":"aa9441a2-a8c5-4f10-89c6-88fa689e9f7c","Type":"ContainerStarted","Data":"c553f1f843772d46f0f1dc032a7e7e48bc4a10b050b9fe64a58cbf3cd775adf2"} Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.253005 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vj8ww"] Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.257039 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ljhpv"] Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.275648 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:58 crc kubenswrapper[4685]: E1013 08:46:58.277887 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:46:58.77783188 +0000 UTC m=+143.925707641 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.279752 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-cnh25"] Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.287072 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" podStartSLOduration=122.287043093 podStartE2EDuration="2m2.287043093s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:58.285445706 +0000 UTC m=+143.433321467" watchObservedRunningTime="2025-10-13 08:46:58.287043093 +0000 UTC m=+143.434918854" Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.287511 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-csgb2" event={"ID":"e8d90978-a395-4313-bcb2-ca35a303b518","Type":"ContainerStarted","Data":"f5210dca9ec30fc068bb330714fc679670486c77154156796244655d4a397ae4"} Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.332627 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-9cnpf" event={"ID":"bcca1906-64bf-44dc-9307-b8d44d2b506f","Type":"ContainerStarted","Data":"43f3d8a52fe5d32fd5564e9cf2e5a317387fa3a61c8886499c331fac68ca572b"} Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.346105 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6v84k" event={"ID":"3937dccc-95e5-4aa3-be81-6d12c31f7cb9","Type":"ContainerStarted","Data":"5fe0fb46edb1b128f4959eaca0a07dc0cfb65b5db116fbd1938a70df731a4943"} Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.371885 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rx99b" event={"ID":"b9f835b0-16cf-4f05-a8a2-2767a929ca25","Type":"ContainerStarted","Data":"e800b413023d44194bd5c565f4acb79cbbfe3f2707f768160a4f2e5499e00360"} Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.372310 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" podStartSLOduration=122.372281237 podStartE2EDuration="2m2.372281237s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:58.372236106 +0000 UTC m=+143.520111877" watchObservedRunningTime="2025-10-13 08:46:58.372281237 +0000 UTC m=+143.520156998" Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.380301 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:46:58 crc kubenswrapper[4685]: E1013 08:46:58.381682 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:46:58.881659644 +0000 UTC m=+144.029535405 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.414469 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" event={"ID":"b57dcc9c-6002-4c47-ae85-c287659cb679","Type":"ContainerStarted","Data":"9a4ad817923ce2d804e10f2987a99288a4a98035f77102286effd7bcae9c150a"} Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.439530 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-8ks9s" event={"ID":"83d9c4d8-10b5-4d1e-a71e-e6e92efd8379","Type":"ContainerStarted","Data":"b93b38bd38358ac5d6d87b18e7d7fb373c5f6a7e50e2470c656daa6a9d1c239a"} Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.443523 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-8ks9s" Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.448837 4685 patch_prober.go:28] interesting pod/console-operator-58897d9998-8ks9s container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.448931 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wn86t"] Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.448948 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-8ks9s" podUID="83d9c4d8-10b5-4d1e-a71e-e6e92efd8379" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.481926 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:58 crc kubenswrapper[4685]: E1013 08:46:58.484470 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:46:58.984453398 +0000 UTC m=+144.132329159 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.495870 4685 generic.go:334] "Generic (PLEG): container finished" podID="2f11d8d5-7c31-42a1-83e5-cfbfa7de648d" containerID="2305ab0afe011e9528cdd2d3d68014e17336f7ba77f1eac56a2788ef572e433c" exitCode=0 Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.496275 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" event={"ID":"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d","Type":"ContainerDied","Data":"2305ab0afe011e9528cdd2d3d68014e17336f7ba77f1eac56a2788ef572e433c"} Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.519280 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pk75f" podStartSLOduration=122.519252478 podStartE2EDuration="2m2.519252478s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:58.460589281 +0000 UTC m=+143.608465042" watchObservedRunningTime="2025-10-13 08:46:58.519252478 +0000 UTC m=+143.667128239" Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.523797 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" event={"ID":"b2158024-4bf6-457d-809d-c389b8ff2309","Type":"ContainerStarted","Data":"53c78249cb47a22e9d6b54303c1e84c19d3830dcd507d0e79734d799ca546eca"} Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.523865 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" event={"ID":"b2158024-4bf6-457d-809d-c389b8ff2309","Type":"ContainerStarted","Data":"612cca0a022e01375aed7ba5cf010b2ccef81567dd9db1a92f3e0ca7ed53c94a"} Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.527056 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.547065 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj"] Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.584106 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:46:58 crc kubenswrapper[4685]: E1013 08:46:58.594359 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:46:59.094330561 +0000 UTC m=+144.242206322 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.597835 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-r96rq" event={"ID":"0fe547bf-ea91-4546-a886-c613abeae02a","Type":"ContainerStarted","Data":"7f3143393109f9107ae7a1ad724a68c8e6e5410e38eb8b35fdc669c53d037d92"} Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.643675 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-n8hmz"] Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.646386 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-2kbsj" event={"ID":"44a7f769-9a31-4555-be16-51e310ac34e3","Type":"ContainerStarted","Data":"679471a05c81bb44960f968dfb887dac67c8537c5ba9a234341bb96f3db97877"} Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.655518 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-2kbsj" Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.714318 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.714879 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.716697 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.716789 4685 patch_prober.go:28] interesting pod/downloads-7954f5f757-2kbsj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.716831 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2kbsj" podUID="44a7f769-9a31-4555-be16-51e310ac34e3" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 13 08:46:58 crc kubenswrapper[4685]: E1013 08:46:58.744822 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:46:59.244799286 +0000 UTC m=+144.392675047 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.765135 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-gx6qj" podStartSLOduration=122.765113577 podStartE2EDuration="2m2.765113577s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:58.761453559 +0000 UTC m=+143.909329320" watchObservedRunningTime="2025-10-13 08:46:58.765113577 +0000 UTC m=+143.912989338" Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.844633 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:46:58 crc kubenswrapper[4685]: E1013 08:46:58.846729 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:46:59.346710463 +0000 UTC m=+144.494586224 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.872503 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6v84k" podStartSLOduration=122.872487166 podStartE2EDuration="2m2.872487166s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:58.871245859 +0000 UTC m=+144.019121610" watchObservedRunningTime="2025-10-13 08:46:58.872487166 +0000 UTC m=+144.020362927" Oct 13 08:46:58 crc kubenswrapper[4685]: W1013 08:46:58.948407 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod991a9129_37d1_4407_a9aa_a675512ec8f1.slice/crio-e5aa33c7043bd8670319e5484cb2b306cdc1baddc6cd7996ab16351b8fe5bc4b WatchSource:0}: Error finding container e5aa33c7043bd8670319e5484cb2b306cdc1baddc6cd7996ab16351b8fe5bc4b: Status 404 returned error can't find the container with id e5aa33c7043bd8670319e5484cb2b306cdc1baddc6cd7996ab16351b8fe5bc4b Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.949422 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:58 crc kubenswrapper[4685]: E1013 08:46:58.949838 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:46:59.449825856 +0000 UTC m=+144.597701617 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:58 crc kubenswrapper[4685]: I1013 08:46:58.970560 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4zvdw"] Oct 13 08:46:59 crc kubenswrapper[4685]: W1013 08:46:59.010637 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podacb21b38_5eb2_4db3_a8ba_9c022c9ae79e.slice/crio-b0642940f2c6a22126028f0024692c36c473b8b709b6fefb36fba952ac510607 WatchSource:0}: Error finding container b0642940f2c6a22126028f0024692c36c473b8b709b6fefb36fba952ac510607: Status 404 returned error can't find the container with id b0642940f2c6a22126028f0024692c36c473b8b709b6fefb36fba952ac510607 Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.051366 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:46:59 crc kubenswrapper[4685]: E1013 08:46:59.051516 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:46:59.551494325 +0000 UTC m=+144.699370086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.052155 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:59 crc kubenswrapper[4685]: E1013 08:46:59.052505 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:46:59.552495035 +0000 UTC m=+144.700370796 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.099056 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-csgb2" podStartSLOduration=123.099037043 podStartE2EDuration="2m3.099037043s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:59.098567099 +0000 UTC m=+144.246442870" watchObservedRunningTime="2025-10-13 08:46:59.099037043 +0000 UTC m=+144.246912804" Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.153798 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:46:59 crc kubenswrapper[4685]: E1013 08:46:59.154360 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:46:59.65433852 +0000 UTC m=+144.802214281 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.196377 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-l74cd" podStartSLOduration=123.196345304 podStartE2EDuration="2m3.196345304s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:59.195844089 +0000 UTC m=+144.343719850" watchObservedRunningTime="2025-10-13 08:46:59.196345304 +0000 UTC m=+144.344221065" Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.261869 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:59 crc kubenswrapper[4685]: E1013 08:46:59.262321 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:46:59.762303786 +0000 UTC m=+144.910179547 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.371834 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k2lwz"] Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.416741 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:46:59 crc kubenswrapper[4685]: E1013 08:46:59.417668 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:46:59.917640226 +0000 UTC m=+145.065515987 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.511964 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" podStartSLOduration=123.511942957 podStartE2EDuration="2m3.511942957s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:59.436726371 +0000 UTC m=+144.584602142" watchObservedRunningTime="2025-10-13 08:46:59.511942957 +0000 UTC m=+144.659818708" Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.547647 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-8ks9s" podStartSLOduration=123.547624274 podStartE2EDuration="2m3.547624274s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:59.516116551 +0000 UTC m=+144.663992322" watchObservedRunningTime="2025-10-13 08:46:59.547624274 +0000 UTC m=+144.695500045" Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.555730 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:59 crc kubenswrapper[4685]: E1013 08:46:59.556087 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:00.056072154 +0000 UTC m=+145.203947915 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.556143 4685 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-ssk8b container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.556164 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" podUID="b2158024-4bf6-457d-809d-c389b8ff2309" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.576301 4685 patch_prober.go:28] interesting pod/router-default-5444994796-r96rq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 08:46:59 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Oct 13 08:46:59 crc kubenswrapper[4685]: [+]process-running ok Oct 13 08:46:59 crc kubenswrapper[4685]: healthz check failed Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.576362 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r96rq" podUID="0fe547bf-ea91-4546-a886-c613abeae02a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.592531 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-r96rq" Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.592586 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xd6tf"] Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.632506 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gpr4j" podStartSLOduration=123.632491057 podStartE2EDuration="2m3.632491057s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:59.630601821 +0000 UTC m=+144.778477582" watchObservedRunningTime="2025-10-13 08:46:59.632491057 +0000 UTC m=+144.780366818" Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.656615 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:46:59 crc kubenswrapper[4685]: E1013 08:46:59.657121 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:00.157103465 +0000 UTC m=+145.304979226 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.673491 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-2kbsj" podStartSLOduration=123.67347441 podStartE2EDuration="2m3.67347441s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:59.672595384 +0000 UTC m=+144.820471145" watchObservedRunningTime="2025-10-13 08:46:59.67347441 +0000 UTC m=+144.821350171" Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.720225 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-r96rq" podStartSLOduration=123.720197283 podStartE2EDuration="2m3.720197283s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:59.717668328 +0000 UTC m=+144.865544099" watchObservedRunningTime="2025-10-13 08:46:59.720197283 +0000 UTC m=+144.868073044" Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.738883 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj" event={"ID":"926501e1-d68a-4c86-b59e-f87c43c3a9ae","Type":"ContainerStarted","Data":"a65c81389824a895c452d22cd61a8a57b6491c2b67c73c8cc90d08ac95e73b37"} Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.764935 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c4d9" event={"ID":"916530ba-091f-4181-aee2-eaa5cd546db5","Type":"ContainerStarted","Data":"6aefa3bd2c1257ce580c3df754a5aa153028880fc45ebb967ec9fd03ce127cfc"} Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.768137 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9wxlv"] Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.779877 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:59 crc kubenswrapper[4685]: E1013 08:46:59.780226 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:00.28021142 +0000 UTC m=+145.428087191 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.840436 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pbpvq" event={"ID":"c2f79e15-fea6-4391-a711-53ca225b1857","Type":"ContainerStarted","Data":"ca2fceaf066217d95a11dc3ea82ec265680a5c3c6aa64da3617944a65cd1ad3a"} Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.892613 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:46:59 crc kubenswrapper[4685]: E1013 08:46:59.894694 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:00.394672189 +0000 UTC m=+145.542548140 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.975160 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn86t" event={"ID":"991a9129-37d1-4407-a9aa-a675512ec8f1","Type":"ContainerStarted","Data":"e5aa33c7043bd8670319e5484cb2b306cdc1baddc6cd7996ab16351b8fe5bc4b"} Oct 13 08:46:59 crc kubenswrapper[4685]: I1013 08:46:59.993965 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:46:59 crc kubenswrapper[4685]: E1013 08:46:59.995569 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:00.495556195 +0000 UTC m=+145.643431956 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.007712 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" event={"ID":"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0","Type":"ContainerStarted","Data":"48f11788ef5cd3f9f1d47877d0279abe11bcb7e7ad8661aaad41270e74ff00a4"} Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.055250 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n8hmz" event={"ID":"acb21b38-5eb2-4db3-a8ba-9c022c9ae79e","Type":"ContainerStarted","Data":"b0642940f2c6a22126028f0024692c36c473b8b709b6fefb36fba952ac510607"} Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.079061 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pbpvq" podStartSLOduration=124.079046047 podStartE2EDuration="2m4.079046047s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:46:59.927843901 +0000 UTC m=+145.075719662" watchObservedRunningTime="2025-10-13 08:47:00.079046047 +0000 UTC m=+145.226921808" Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.099859 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:00 crc kubenswrapper[4685]: E1013 08:47:00.100552 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:00.600526384 +0000 UTC m=+145.748402145 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.135278 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8x7bf" event={"ID":"e6e5e0a0-6ff8-44eb-99f6-eb9a4eb2aa47","Type":"ContainerStarted","Data":"fc644484a2cfd85ed413f4924f52b5e5e625edad8ce092fa1b93282cd44aa20e"} Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.135350 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8x7bf" event={"ID":"e6e5e0a0-6ff8-44eb-99f6-eb9a4eb2aa47","Type":"ContainerStarted","Data":"fe11eaccf45cea293fd939cf7cd413ec74eb5402770d1977005c985aca97602c"} Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.162610 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-9cnpf" event={"ID":"bcca1906-64bf-44dc-9307-b8d44d2b506f","Type":"ContainerStarted","Data":"73c20d41ba43203391fa56139366878d0d4b86df6aa7877220851763183e264c"} Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.183203 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-cnh25" event={"ID":"53a5a736-ebd0-4c09-90b3-d70a4943c038","Type":"ContainerStarted","Data":"ddf2b9f4c629d9af0082409897d53bb9d873cc9852e86bfca623e6c7c3d95f6b"} Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.202477 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:00 crc kubenswrapper[4685]: E1013 08:47:00.204034 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:00.704015348 +0000 UTC m=+145.851891309 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.219668 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8x7bf" podStartSLOduration=124.21964337 podStartE2EDuration="2m4.21964337s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:00.218161716 +0000 UTC m=+145.366037477" watchObservedRunningTime="2025-10-13 08:47:00.21964337 +0000 UTC m=+145.367519131" Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.219807 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-rx99b" podStartSLOduration=124.219800714 podStartE2EDuration="2m4.219800714s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:00.119832154 +0000 UTC m=+145.267707915" watchObservedRunningTime="2025-10-13 08:47:00.219800714 +0000 UTC m=+145.367676485" Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.232316 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-wmjcx" event={"ID":"6e73973b-70de-449a-9ad2-b965b7a83e1f","Type":"ContainerStarted","Data":"f6673945323305feccbed8767ae56e99bdd736aa9a586879b2697b94d67dee7b"} Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.232379 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-wmjcx" event={"ID":"6e73973b-70de-449a-9ad2-b965b7a83e1f","Type":"ContainerStarted","Data":"3118040bb76cd3523d2f71f2d6ed9b1c39b4831e763e49e95de403da4a374617"} Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.233831 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ljhpv" event={"ID":"9c926cbb-e6ad-4e49-9a6c-afaf6eb9ced6","Type":"ContainerStarted","Data":"9d93ccd32a360cb42a10663cce56ae6dcbb81ffc3b9c00e0cf7ffc9563a1c3e9"} Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.265543 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-ps67b"] Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.277509 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4zvdw" event={"ID":"d869fa03-5196-4a23-a7d3-9bb709891678","Type":"ContainerStarted","Data":"abeda7508370ade805d51007301d052863cedf026d35d1c0a11d5b21a98cd603"} Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.313409 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2bwvx" event={"ID":"4ef87f11-a2ee-48e5-86ec-94283d235933","Type":"ContainerStarted","Data":"baaf71cbb022632bdcb1448ea34556e6b87fa927aa5f081cc01c316c2c13d2eb"} Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.314467 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:00 crc kubenswrapper[4685]: E1013 08:47:00.314670 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:00.814641532 +0000 UTC m=+145.962517293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.314797 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:00 crc kubenswrapper[4685]: E1013 08:47:00.315195 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:00.815181278 +0000 UTC m=+145.963057039 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.372304 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7nbg"] Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.376314 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-lcljr" event={"ID":"c96407ea-0763-479d-8020-ace1607bd427","Type":"ContainerStarted","Data":"d2d45343e0b64969a9a9f283c8b5643d98d6f343760333bdab2a7c749ea24bed"} Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.386993 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vj8ww" event={"ID":"c1eea5fd-9b4d-4d4d-a715-6523309cdd87","Type":"ContainerStarted","Data":"0eb54243688d454920816c37dd7e623e14aea12c3907c9d9947d10c0fc63d581"} Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.401658 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j6kt6" event={"ID":"a185671e-0a3a-4a9e-a884-65b448b1e922","Type":"ContainerStarted","Data":"f4eb2ab013e818137dea9e9d7cd3d731109dc438bfc9fb1bf01bd8e7427d5ff0"} Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.416017 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:00 crc kubenswrapper[4685]: E1013 08:47:00.417248 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:00.91723246 +0000 UTC m=+146.065108211 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.425787 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k2lwz" event={"ID":"571d990f-04d9-4e62-902b-f44b7b666901","Type":"ContainerStarted","Data":"723291a0f443b753e0fa4a930f46787221cf38dd47720f8f5e23aa292233bd4f"} Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.454388 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84w94" event={"ID":"30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90","Type":"ContainerStarted","Data":"cab1a8cb3df5f6a54837ff7088965f99d809947bf56cbc0202faedbf47b7e794"} Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.454446 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84w94" event={"ID":"30e776cf-ad99-4bdf-bd89-e8fdf7eb8b90","Type":"ContainerStarted","Data":"4b2f1509c99264eb15b7c4114ab4fe19aedfdf3d4e028fc04c55abe76027cb91"} Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.504822 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-csgb2" event={"ID":"e8d90978-a395-4313-bcb2-ca35a303b518","Type":"ContainerStarted","Data":"8a09af0877137aa1ad4644c827f9097e30a8e1d2a011ad3b959ef296b2d5c8ba"} Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.521650 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.523193 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-96w58"] Oct 13 08:47:00 crc kubenswrapper[4685]: E1013 08:47:00.523608 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:01.023583639 +0000 UTC m=+146.171459400 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.559110 4685 patch_prober.go:28] interesting pod/router-default-5444994796-r96rq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 08:47:00 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Oct 13 08:47:00 crc kubenswrapper[4685]: [+]process-running ok Oct 13 08:47:00 crc kubenswrapper[4685]: healthz check failed Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.559204 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r96rq" podUID="0fe547bf-ea91-4546-a886-c613abeae02a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.625623 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:00 crc kubenswrapper[4685]: E1013 08:47:00.626551 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:01.126504125 +0000 UTC m=+146.274379886 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.636760 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn" event={"ID":"3672a087-9d9b-45ee-8f80-686959b395d6","Type":"ContainerStarted","Data":"36e26e5c5a3fa5ed680abafb9f41cc8050b80f6b5327664df8fa08c6b89d51a6"} Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.677288 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-9cnpf" podStartSLOduration=124.677261198 podStartE2EDuration="2m4.677261198s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:00.519327372 +0000 UTC m=+145.667203143" watchObservedRunningTime="2025-10-13 08:47:00.677261198 +0000 UTC m=+145.825136959" Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.706118 4685 generic.go:334] "Generic (PLEG): container finished" podID="aa9441a2-a8c5-4f10-89c6-88fa689e9f7c" containerID="f5029de93c82b252ac0a3aefc952cd79e1cb373dafbd41a5a26e948a4346b5e8" exitCode=0 Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.706693 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-j6kt6" podStartSLOduration=124.706660478 podStartE2EDuration="2m4.706660478s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:00.674366682 +0000 UTC m=+145.822242443" watchObservedRunningTime="2025-10-13 08:47:00.706660478 +0000 UTC m=+145.854536239" Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.710682 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4lflr" event={"ID":"aa9441a2-a8c5-4f10-89c6-88fa689e9f7c","Type":"ContainerDied","Data":"f5029de93c82b252ac0a3aefc952cd79e1cb373dafbd41a5a26e948a4346b5e8"} Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.710734 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7sx5f"] Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.710796 4685 patch_prober.go:28] interesting pod/downloads-7954f5f757-2kbsj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.710825 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2kbsj" podUID="44a7f769-9a31-4555-be16-51e310ac34e3" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.724002 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.727435 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:00 crc kubenswrapper[4685]: E1013 08:47:00.729030 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:01.22900699 +0000 UTC m=+146.376882751 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.763891 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-lcljr" podStartSLOduration=124.763867822 podStartE2EDuration="2m4.763867822s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:00.763416799 +0000 UTC m=+145.911292550" watchObservedRunningTime="2025-10-13 08:47:00.763867822 +0000 UTC m=+145.911743583" Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.829427 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:00 crc kubenswrapper[4685]: E1013 08:47:00.831872 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:01.331857235 +0000 UTC m=+146.479732996 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.872604 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-wmjcx" podStartSLOduration=7.87257003 podStartE2EDuration="7.87257003s" podCreationTimestamp="2025-10-13 08:46:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:00.851766205 +0000 UTC m=+145.999641966" watchObservedRunningTime="2025-10-13 08:47:00.87257003 +0000 UTC m=+146.020445801" Oct 13 08:47:00 crc kubenswrapper[4685]: I1013 08:47:00.946892 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:00 crc kubenswrapper[4685]: E1013 08:47:00.947408 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:01.447384185 +0000 UTC m=+146.595259946 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.008593 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84w94" podStartSLOduration=125.008576087 podStartE2EDuration="2m5.008576087s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:00.933036251 +0000 UTC m=+146.080912012" watchObservedRunningTime="2025-10-13 08:47:01.008576087 +0000 UTC m=+146.156451848" Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.047553 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:01 crc kubenswrapper[4685]: E1013 08:47:01.048083 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:01.548064516 +0000 UTC m=+146.695940277 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.149890 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:01 crc kubenswrapper[4685]: E1013 08:47:01.150266 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:01.650252011 +0000 UTC m=+146.798127772 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.169556 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn" podStartSLOduration=121.169534362 podStartE2EDuration="2m1.169534362s" podCreationTimestamp="2025-10-13 08:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:01.081346772 +0000 UTC m=+146.229222533" watchObservedRunningTime="2025-10-13 08:47:01.169534362 +0000 UTC m=+146.317410113" Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.252094 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:01 crc kubenswrapper[4685]: E1013 08:47:01.252285 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:01.752260562 +0000 UTC m=+146.900136323 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.252329 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:01 crc kubenswrapper[4685]: E1013 08:47:01.252710 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:01.752701195 +0000 UTC m=+146.900576956 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.353623 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:01 crc kubenswrapper[4685]: E1013 08:47:01.354178 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:01.854160208 +0000 UTC m=+147.002035959 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.354596 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:01 crc kubenswrapper[4685]: E1013 08:47:01.355016 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:01.855007333 +0000 UTC m=+147.002883094 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.455444 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:01 crc kubenswrapper[4685]: E1013 08:47:01.455943 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:01.95590741 +0000 UTC m=+147.103783171 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.495874 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.496571 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.509058 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.509575 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.570955 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.573434 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.573694 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4e31f8b2-2961-4bb6-9465-9d2e66c9f000-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4e31f8b2-2961-4bb6-9465-9d2e66c9f000\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.573889 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4e31f8b2-2961-4bb6-9465-9d2e66c9f000-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4e31f8b2-2961-4bb6-9465-9d2e66c9f000\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 13 08:47:01 crc kubenswrapper[4685]: E1013 08:47:01.575513 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:02.075486611 +0000 UTC m=+147.223362372 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.585777 4685 patch_prober.go:28] interesting pod/router-default-5444994796-r96rq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 08:47:01 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Oct 13 08:47:01 crc kubenswrapper[4685]: [+]process-running ok Oct 13 08:47:01 crc kubenswrapper[4685]: healthz check failed Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.585835 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r96rq" podUID="0fe547bf-ea91-4546-a886-c613abeae02a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.675905 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:01 crc kubenswrapper[4685]: E1013 08:47:01.676278 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:02.176250784 +0000 UTC m=+147.324126545 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.676387 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4e31f8b2-2961-4bb6-9465-9d2e66c9f000-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4e31f8b2-2961-4bb6-9465-9d2e66c9f000\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.676476 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4e31f8b2-2961-4bb6-9465-9d2e66c9f000-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4e31f8b2-2961-4bb6-9465-9d2e66c9f000\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.676501 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:01 crc kubenswrapper[4685]: E1013 08:47:01.676860 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:02.176850222 +0000 UTC m=+147.324725993 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.677451 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4e31f8b2-2961-4bb6-9465-9d2e66c9f000-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4e31f8b2-2961-4bb6-9465-9d2e66c9f000\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.723637 4685 patch_prober.go:28] interesting pod/console-operator-58897d9998-8ks9s container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.723687 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-8ks9s" podUID="83d9c4d8-10b5-4d1e-a71e-e6e92efd8379" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.739187 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4e31f8b2-2961-4bb6-9465-9d2e66c9f000-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4e31f8b2-2961-4bb6-9465-9d2e66c9f000\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.766783 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" event={"ID":"2f11d8d5-7c31-42a1-83e5-cfbfa7de648d","Type":"ContainerStarted","Data":"d42b1747703188d7f8ed59125d2c169ccae90ae623b01f04f5f4d80bb75c2227"} Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.769216 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-96w58" event={"ID":"763388db-ca9e-432e-9c91-4821392af1a1","Type":"ContainerStarted","Data":"feda996eb4b133bed39073cb0f5dd15c286bcfe02e4898dc48c8b3f210125337"} Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.777193 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:01 crc kubenswrapper[4685]: E1013 08:47:01.777558 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:02.277539513 +0000 UTC m=+147.425415274 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.784564 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ljhpv" event={"ID":"9c926cbb-e6ad-4e49-9a6c-afaf6eb9ced6","Type":"ContainerStarted","Data":"51b674c1f4af6b2c0b08abcd1ecd48e576e139c899d6060d89f3b1288199130f"} Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.800060 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" podStartSLOduration=125.800025338 podStartE2EDuration="2m5.800025338s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:01.798656278 +0000 UTC m=+146.946532039" watchObservedRunningTime="2025-10-13 08:47:01.800025338 +0000 UTC m=+146.947901099" Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.815697 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-cnh25" event={"ID":"53a5a736-ebd0-4c09-90b3-d70a4943c038","Type":"ContainerStarted","Data":"7456024b7c9f981fc49b6ac53438471a7cfd08611369728bc5477dbb0ca18b30"} Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.831489 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rx99b" event={"ID":"b9f835b0-16cf-4f05-a8a2-2767a929ca25","Type":"ContainerStarted","Data":"5d8b5e1e36734d9fc5c1a0bc107e0685670470b829c1637201a2117e38f0be56"} Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.832400 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.855157 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2bwvx" event={"ID":"4ef87f11-a2ee-48e5-86ec-94283d235933","Type":"ContainerStarted","Data":"a344e8d099535b8cdba38f20b8ea7430f62f51238aa90d7dfa985e687656d0c1"} Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.862741 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7nbg" event={"ID":"bd767915-7aeb-45a4-96a3-c2f636c5e754","Type":"ContainerStarted","Data":"fc829c11bab96839b173a22652541dd43c441c9730f9006714ddc1a8676ef24b"} Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.862791 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7nbg" event={"ID":"bd767915-7aeb-45a4-96a3-c2f636c5e754","Type":"ContainerStarted","Data":"7df1e169161c16994370f1de16049e827d72a41b307d08c915097b681b60fc44"} Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.879216 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:01 crc kubenswrapper[4685]: E1013 08:47:01.885152 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:02.385132449 +0000 UTC m=+147.533008310 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.897835 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c4d9" event={"ID":"916530ba-091f-4181-aee2-eaa5cd546db5","Type":"ContainerStarted","Data":"68e4c388e4f69544b7ee712a2173152cdd029da3ee8d266d8fa96136e9733d5d"} Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.897882 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c4d9" event={"ID":"916530ba-091f-4181-aee2-eaa5cd546db5","Type":"ContainerStarted","Data":"71866db3e985748f9eb256fb54ce93a6e0f8dbb14d9eb4ffa8e99baafafc24a0"} Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.945525 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj" event={"ID":"926501e1-d68a-4c86-b59e-f87c43c3a9ae","Type":"ContainerStarted","Data":"919d906bf07e6a30255d10ab418448430ff6626368800a053e227ce76e13038a"} Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.947034 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj" Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.957896 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2bwvx" podStartSLOduration=125.957874842 podStartE2EDuration="2m5.957874842s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:01.956991326 +0000 UTC m=+147.104867097" watchObservedRunningTime="2025-10-13 08:47:01.957874842 +0000 UTC m=+147.105750603" Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.958938 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ljhpv" podStartSLOduration=125.958911793 podStartE2EDuration="2m5.958911793s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:01.870356031 +0000 UTC m=+147.018231812" watchObservedRunningTime="2025-10-13 08:47:01.958911793 +0000 UTC m=+147.106787554" Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.959938 4685 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-vh7kj container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.28:5443/healthz\": dial tcp 10.217.0.28:5443: connect: connection refused" start-of-body= Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.959973 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj" podUID="926501e1-d68a-4c86-b59e-f87c43c3a9ae" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.28:5443/healthz\": dial tcp 10.217.0.28:5443: connect: connection refused" Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.976586 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4zvdw" event={"ID":"d869fa03-5196-4a23-a7d3-9bb709891678","Type":"ContainerStarted","Data":"34dc7ccc4fe292333e7f7cbbfa50bb5f4203d25e23033c2b651a821a54978a99"} Oct 13 08:47:01 crc kubenswrapper[4685]: I1013 08:47:01.987114 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:01 crc kubenswrapper[4685]: E1013 08:47:01.988942 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:02.488897191 +0000 UTC m=+147.636772952 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.000632 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn" event={"ID":"3672a087-9d9b-45ee-8f80-686959b395d6","Type":"ContainerStarted","Data":"c41f5b326a5ed917fc98da2be46741a8baa16dd62bb2a2fb843af8060340137f"} Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.005691 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn86t" event={"ID":"991a9129-37d1-4407-a9aa-a675512ec8f1","Type":"ContainerStarted","Data":"e08a1185be184c94ca3c4c17e6a1edcf34dc3023702dbc93513c867d22140b36"} Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.016423 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c4d9" podStartSLOduration=126.016402664 podStartE2EDuration="2m6.016402664s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:02.011481409 +0000 UTC m=+147.159357170" watchObservedRunningTime="2025-10-13 08:47:02.016402664 +0000 UTC m=+147.164278425" Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.048936 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vj8ww" event={"ID":"c1eea5fd-9b4d-4d4d-a715-6523309cdd87","Type":"ContainerStarted","Data":"716c6e394a522c67d5331a94a51ddbbe93e98abded60809e193ed9edeaacfdb7"} Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.050209 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vj8ww" Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.050231 4685 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-vj8ww container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.050360 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vj8ww" podUID="c1eea5fd-9b4d-4d4d-a715-6523309cdd87" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.082785 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k2lwz" event={"ID":"571d990f-04d9-4e62-902b-f44b7b666901","Type":"ContainerStarted","Data":"9666e623efae93faabe35e2be7a952c8bb43728f4c1c43f5789aca01781a2d9a"} Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.084438 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k2lwz" Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.086274 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj" podStartSLOduration=126.086236852 podStartE2EDuration="2m6.086236852s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:02.069705783 +0000 UTC m=+147.217581564" watchObservedRunningTime="2025-10-13 08:47:02.086236852 +0000 UTC m=+147.234112613" Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.092212 4685 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-k2lwz container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.092289 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k2lwz" podUID="571d990f-04d9-4e62-902b-f44b7b666901" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.092332 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:02 crc kubenswrapper[4685]: E1013 08:47:02.094555 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:02.594536658 +0000 UTC m=+147.742412419 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.102010 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn86t" podStartSLOduration=126.101983088 podStartE2EDuration="2m6.101983088s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:02.100619989 +0000 UTC m=+147.248495750" watchObservedRunningTime="2025-10-13 08:47:02.101983088 +0000 UTC m=+147.249858849" Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.112683 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" event={"ID":"b8ae1ce1-bf5f-4d13-87b4-0810f63434af","Type":"ContainerStarted","Data":"e961755b572021234363fccf249332fc4b3c12ce669dfb40dba16f9676a3689e"} Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.113865 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.115021 4685 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7sx5f container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.115083 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" podUID="b8ae1ce1-bf5f-4d13-87b4-0810f63434af" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.154582 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" event={"ID":"8f247d85-48fa-49a1-9c0c-3ec7b7d202c0","Type":"ContainerStarted","Data":"0d4faede8d8cd70fc48831f36e8f31d97705c8797229b78ef75caf5d1e7e3b33"} Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.178063 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-9wxlv" event={"ID":"51d519c2-0060-4aa2-9b25-11c742aca2d1","Type":"ContainerStarted","Data":"264e2c3dc338cfe487817bf16edbf03753c940e3a47ef4be6ec8b28b70e3fbc3"} Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.178114 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-9wxlv" event={"ID":"51d519c2-0060-4aa2-9b25-11c742aca2d1","Type":"ContainerStarted","Data":"b7bad7b325bbf5a03a9ec0c71f1a55de1939a5327a0404cc96c85ecea0fb2c5a"} Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.189571 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4zvdw" podStartSLOduration=126.183748919 podStartE2EDuration="2m6.183748919s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:02.181335857 +0000 UTC m=+147.329211608" watchObservedRunningTime="2025-10-13 08:47:02.183748919 +0000 UTC m=+147.331624680" Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.195590 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:02 crc kubenswrapper[4685]: E1013 08:47:02.196640 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:02.69661602 +0000 UTC m=+147.844491781 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.203484 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n8hmz" event={"ID":"acb21b38-5eb2-4db3-a8ba-9c022c9ae79e","Type":"ContainerStarted","Data":"7bead96361c247534afdab406c191aae105dca0f4989f71947d64bb92e36b748"} Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.203541 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n8hmz" event={"ID":"acb21b38-5eb2-4db3-a8ba-9c022c9ae79e","Type":"ContainerStarted","Data":"29569da5dafa4a0c7741b739613b154ed1825affa72a7874ce620dced0f8b689"} Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.233438 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-ps67b" event={"ID":"9ab91fd8-27d8-458a-97b9-09b9ba008d67","Type":"ContainerStarted","Data":"affb90173377af81dd0da867f61032b96a69f23262098d5f572e986121134e1c"} Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.233485 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-ps67b" event={"ID":"9ab91fd8-27d8-458a-97b9-09b9ba008d67","Type":"ContainerStarted","Data":"e01cb53ff4cd9b5c870b6a11ad40d0b40d4d2e2a0e440a1e7ab12de4548bbf18"} Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.248333 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xd6tf" event={"ID":"47f6f699-d6e2-4289-bd8a-9ec387887523","Type":"ContainerStarted","Data":"32c1050ec9abd6a54d2bc996d186032b8ed6e2322fc41718eee20c44edb5189f"} Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.248382 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xd6tf" event={"ID":"47f6f699-d6e2-4289-bd8a-9ec387887523","Type":"ContainerStarted","Data":"b4f07176c2212bb1502ceda226f6e4bc61663f1ddddeb8c857957d5e1661057c"} Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.297648 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:02 crc kubenswrapper[4685]: E1013 08:47:02.301068 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:02.801054142 +0000 UTC m=+147.948929903 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.399133 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:02 crc kubenswrapper[4685]: E1013 08:47:02.399745 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:02.899718463 +0000 UTC m=+148.047594224 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.480789 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" podStartSLOduration=126.480755362 podStartE2EDuration="2m6.480755362s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:02.478579718 +0000 UTC m=+147.626455479" watchObservedRunningTime="2025-10-13 08:47:02.480755362 +0000 UTC m=+147.628631133" Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.483699 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vj8ww" podStartSLOduration=126.483685399 podStartE2EDuration="2m6.483685399s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:02.249444865 +0000 UTC m=+147.397320646" watchObservedRunningTime="2025-10-13 08:47:02.483685399 +0000 UTC m=+147.631561170" Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.501360 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:02 crc kubenswrapper[4685]: E1013 08:47:02.501691 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:03.001680692 +0000 UTC m=+148.149556453 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.559518 4685 patch_prober.go:28] interesting pod/router-default-5444994796-r96rq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 08:47:02 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Oct 13 08:47:02 crc kubenswrapper[4685]: [+]process-running ok Oct 13 08:47:02 crc kubenswrapper[4685]: healthz check failed Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.559594 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r96rq" podUID="0fe547bf-ea91-4546-a886-c613abeae02a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.602879 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:02 crc kubenswrapper[4685]: E1013 08:47:02.603458 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:03.103440755 +0000 UTC m=+148.251316516 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.645635 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" podStartSLOduration=126.645617773 podStartE2EDuration="2m6.645617773s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:02.640547243 +0000 UTC m=+147.788423004" watchObservedRunningTime="2025-10-13 08:47:02.645617773 +0000 UTC m=+147.793493534" Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.708355 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:02 crc kubenswrapper[4685]: E1013 08:47:02.708820 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:03.208803544 +0000 UTC m=+148.356679305 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.780845 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k2lwz" podStartSLOduration=126.780820846 podStartE2EDuration="2m6.780820846s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:02.728053144 +0000 UTC m=+147.875928905" watchObservedRunningTime="2025-10-13 08:47:02.780820846 +0000 UTC m=+147.928696617" Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.783247 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-ps67b" podStartSLOduration=9.783236368 podStartE2EDuration="9.783236368s" podCreationTimestamp="2025-10-13 08:46:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:02.779298791 +0000 UTC m=+147.927174552" watchObservedRunningTime="2025-10-13 08:47:02.783236368 +0000 UTC m=+147.931112129" Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.809089 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:02 crc kubenswrapper[4685]: E1013 08:47:02.809357 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:03.30931329 +0000 UTC m=+148.457189051 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.809535 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:02 crc kubenswrapper[4685]: E1013 08:47:02.809981 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:03.309970509 +0000 UTC m=+148.457846460 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:02 crc kubenswrapper[4685]: I1013 08:47:02.911474 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:02 crc kubenswrapper[4685]: E1013 08:47:02.911840 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:03.411795954 +0000 UTC m=+148.559671715 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.013903 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:03 crc kubenswrapper[4685]: E1013 08:47:03.014504 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:03.514474524 +0000 UTC m=+148.662350285 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.021851 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n8hmz" podStartSLOduration=127.021825761 podStartE2EDuration="2m7.021825761s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:02.876755607 +0000 UTC m=+148.024631378" watchObservedRunningTime="2025-10-13 08:47:03.021825761 +0000 UTC m=+148.169701542" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.114799 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:03 crc kubenswrapper[4685]: E1013 08:47:03.114957 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:03.614906308 +0000 UTC m=+148.762782069 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.115430 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:03 crc kubenswrapper[4685]: E1013 08:47:03.115954 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:03.615927807 +0000 UTC m=+148.763803568 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.216632 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:03 crc kubenswrapper[4685]: E1013 08:47:03.216887 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:03.716838265 +0000 UTC m=+148.864714026 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.217181 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:03 crc kubenswrapper[4685]: E1013 08:47:03.217672 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:03.717661869 +0000 UTC m=+148.865537630 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.255203 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-cnh25" event={"ID":"53a5a736-ebd0-4c09-90b3-d70a4943c038","Type":"ContainerStarted","Data":"2454e4b28a383621bbf0d1bfce9801ca2c350eba1a7e623a007ce93acba8b871"} Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.256641 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-cnh25" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.259224 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xd6tf" event={"ID":"47f6f699-d6e2-4289-bd8a-9ec387887523","Type":"ContainerStarted","Data":"5077e4a665895ff68320213dacfa70524118e51973c34ea30e04b76a128e0fbb"} Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.261383 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7nbg" event={"ID":"bd767915-7aeb-45a4-96a3-c2f636c5e754","Type":"ContainerStarted","Data":"097d14d9bffa52bc800285eb0bd25d58f1904e5221f2bf96ff9a285f77ce3a3f"} Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.261504 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7nbg" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.263317 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" event={"ID":"b8ae1ce1-bf5f-4d13-87b4-0810f63434af","Type":"ContainerStarted","Data":"6379ee3dd8ba0c8279c34e7d87be880c84fb191bd775ff824985be478e72968d"} Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.264072 4685 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7sx5f container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.264139 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" podUID="b8ae1ce1-bf5f-4d13-87b4-0810f63434af" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.267235 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4lflr" event={"ID":"aa9441a2-a8c5-4f10-89c6-88fa689e9f7c","Type":"ContainerStarted","Data":"c8020357e267f38169a83862c45abc85c04c69e7b1ced9016d07d8e3a68d4bd4"} Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.267291 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4lflr" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.268583 4685 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-k2lwz container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.268659 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k2lwz" podUID="571d990f-04d9-4e62-902b-f44b7b666901" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.297955 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vj8ww" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.320419 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.320885 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.320981 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:47:03 crc kubenswrapper[4685]: E1013 08:47:03.322113 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:03.822090921 +0000 UTC m=+148.969966672 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.324232 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-9wxlv" podStartSLOduration=127.324204193 podStartE2EDuration="2m7.324204193s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:03.024285274 +0000 UTC m=+148.172161035" watchObservedRunningTime="2025-10-13 08:47:03.324204193 +0000 UTC m=+148.472079954" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.324998 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.326643 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.347708 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.387254 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-cnh25" podStartSLOduration=10.38723131 podStartE2EDuration="10.38723131s" podCreationTimestamp="2025-10-13 08:46:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:03.385365154 +0000 UTC m=+148.533240915" watchObservedRunningTime="2025-10-13 08:47:03.38723131 +0000 UTC m=+148.535107071" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.424186 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.424420 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.424763 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:47:03 crc kubenswrapper[4685]: E1013 08:47:03.430141 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:03.930121619 +0000 UTC m=+149.077997390 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.433176 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.439822 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.463788 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.525520 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:03 crc kubenswrapper[4685]: E1013 08:47:03.526542 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:04.026521984 +0000 UTC m=+149.174397745 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.527157 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:03 crc kubenswrapper[4685]: E1013 08:47:03.527508 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:04.027495442 +0000 UTC m=+149.175371203 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.562526 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4lflr" podStartSLOduration=127.562509788 podStartE2EDuration="2m7.562509788s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:03.558339755 +0000 UTC m=+148.706215516" watchObservedRunningTime="2025-10-13 08:47:03.562509788 +0000 UTC m=+148.710385549" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.566485 4685 patch_prober.go:28] interesting pod/router-default-5444994796-r96rq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 08:47:03 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Oct 13 08:47:03 crc kubenswrapper[4685]: [+]process-running ok Oct 13 08:47:03 crc kubenswrapper[4685]: healthz check failed Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.566884 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r96rq" podUID="0fe547bf-ea91-4546-a886-c613abeae02a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.627062 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.627815 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:03 crc kubenswrapper[4685]: E1013 08:47:03.628125 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:04.128100101 +0000 UTC m=+149.275975862 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.724195 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.729632 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:03 crc kubenswrapper[4685]: E1013 08:47:03.730115 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:04.23010187 +0000 UTC m=+149.377977631 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.769087 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xd6tf" podStartSLOduration=127.769070084 podStartE2EDuration="2m7.769070084s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:03.764337174 +0000 UTC m=+148.912212935" watchObservedRunningTime="2025-10-13 08:47:03.769070084 +0000 UTC m=+148.916945845" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.832960 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:03 crc kubenswrapper[4685]: E1013 08:47:03.833669 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:04.333652256 +0000 UTC m=+149.481528017 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.897657 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.905525 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7nbg" podStartSLOduration=127.905497633 podStartE2EDuration="2m7.905497633s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:03.904595787 +0000 UTC m=+149.052471568" watchObservedRunningTime="2025-10-13 08:47:03.905497633 +0000 UTC m=+149.053373394" Oct 13 08:47:03 crc kubenswrapper[4685]: I1013 08:47:03.935048 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:03 crc kubenswrapper[4685]: E1013 08:47:03.935415 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:04.435403148 +0000 UTC m=+149.583278899 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.036844 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:04 crc kubenswrapper[4685]: E1013 08:47:04.038137 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:04.53811972 +0000 UTC m=+149.685995481 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.140700 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:04 crc kubenswrapper[4685]: E1013 08:47:04.141106 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:04.641094188 +0000 UTC m=+149.788969949 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.241676 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:04 crc kubenswrapper[4685]: E1013 08:47:04.242279 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:04.742258693 +0000 UTC m=+149.890134454 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.271269 4685 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-vh7kj container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.28:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.271349 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj" podUID="926501e1-d68a-4c86-b59e-f87c43c3a9ae" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.28:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.299336 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4e31f8b2-2961-4bb6-9465-9d2e66c9f000","Type":"ContainerStarted","Data":"793f9f083d5c31ff31e1c0ca7d704fdbae10d705348e730aff6a679006402e16"} Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.299403 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4e31f8b2-2961-4bb6-9465-9d2e66c9f000","Type":"ContainerStarted","Data":"ec388fedf392452f1b998dee1ad0b506559cb41e8f7539c1ec383c05089734f2"} Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.311196 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-96w58" event={"ID":"763388db-ca9e-432e-9c91-4821392af1a1","Type":"ContainerStarted","Data":"dd7d784ba4e49f4f9ce0fe70831bbcaa5d54868ab422012c1141b7849089462a"} Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.315339 4685 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7sx5f container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.315382 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" podUID="b8ae1ce1-bf5f-4d13-87b4-0810f63434af" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.338198 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k2lwz" Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.344929 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:04 crc kubenswrapper[4685]: E1013 08:47:04.345351 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:04.845332745 +0000 UTC m=+149.993208506 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.418498 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.41847018 podStartE2EDuration="3.41847018s" podCreationTimestamp="2025-10-13 08:47:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:04.407857096 +0000 UTC m=+149.555732857" watchObservedRunningTime="2025-10-13 08:47:04.41847018 +0000 UTC m=+149.566345941" Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.449285 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:04 crc kubenswrapper[4685]: E1013 08:47:04.450808 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:04.950776626 +0000 UTC m=+150.098652587 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.559752 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:04 crc kubenswrapper[4685]: E1013 08:47:04.560138 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:05.060123725 +0000 UTC m=+150.207999486 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.582545 4685 patch_prober.go:28] interesting pod/router-default-5444994796-r96rq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 08:47:04 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Oct 13 08:47:04 crc kubenswrapper[4685]: [+]process-running ok Oct 13 08:47:04 crc kubenswrapper[4685]: healthz check failed Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.582616 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r96rq" podUID="0fe547bf-ea91-4546-a886-c613abeae02a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.663499 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:04 crc kubenswrapper[4685]: E1013 08:47:04.663871 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:05.163850636 +0000 UTC m=+150.311726397 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.766225 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:04 crc kubenswrapper[4685]: E1013 08:47:04.766690 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:05.266671379 +0000 UTC m=+150.414547140 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.767704 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.767759 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.868663 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:04 crc kubenswrapper[4685]: E1013 08:47:04.870776 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:05.37075298 +0000 UTC m=+150.518628741 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.933166 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.933209 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.949154 4685 patch_prober.go:28] interesting pod/console-f9d7485db-gx6qj container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.949255 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-gx6qj" podUID="16484304-6976-45e8-a495-f1d2ee367f0f" containerName="console" probeResult="failure" output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" Oct 13 08:47:04 crc kubenswrapper[4685]: I1013 08:47:04.969892 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:04 crc kubenswrapper[4685]: E1013 08:47:04.970397 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:05.470347749 +0000 UTC m=+150.618223510 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.023092 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vdbjj"] Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.025105 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vdbjj" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.031242 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.072055 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:05 crc kubenswrapper[4685]: E1013 08:47:05.072495 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:05.572464042 +0000 UTC m=+150.720339803 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.072539 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:05 crc kubenswrapper[4685]: E1013 08:47:05.073986 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:05.573963557 +0000 UTC m=+150.721839318 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.160668 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.160715 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.171610 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vdbjj"] Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.181323 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wbzsb"] Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.184264 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.184547 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ljd6\" (UniqueName: \"kubernetes.io/projected/23545a99-b853-4a78-85a6-6186dc3bf903-kube-api-access-8ljd6\") pod \"community-operators-vdbjj\" (UID: \"23545a99-b853-4a78-85a6-6186dc3bf903\") " pod="openshift-marketplace/community-operators-vdbjj" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.184654 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23545a99-b853-4a78-85a6-6186dc3bf903-utilities\") pod \"community-operators-vdbjj\" (UID: \"23545a99-b853-4a78-85a6-6186dc3bf903\") " pod="openshift-marketplace/community-operators-vdbjj" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.184678 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23545a99-b853-4a78-85a6-6186dc3bf903-catalog-content\") pod \"community-operators-vdbjj\" (UID: \"23545a99-b853-4a78-85a6-6186dc3bf903\") " pod="openshift-marketplace/community-operators-vdbjj" Oct 13 08:47:05 crc kubenswrapper[4685]: E1013 08:47:05.184807 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:05.684784767 +0000 UTC m=+150.832660528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.189677 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wbzsb" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.193565 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.196574 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.219030 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wbzsb"] Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.250131 4685 patch_prober.go:28] interesting pod/downloads-7954f5f757-2kbsj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.250215 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2kbsj" podUID="44a7f769-9a31-4555-be16-51e310ac34e3" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.250629 4685 patch_prober.go:28] interesting pod/downloads-7954f5f757-2kbsj container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.250765 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-2kbsj" podUID="44a7f769-9a31-4555-be16-51e310ac34e3" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.251271 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-8ks9s" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.292720 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ca80dd4-6294-4263-b938-188c17814556-utilities\") pod \"certified-operators-wbzsb\" (UID: \"1ca80dd4-6294-4263-b938-188c17814556\") " pod="openshift-marketplace/certified-operators-wbzsb" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.293249 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23545a99-b853-4a78-85a6-6186dc3bf903-utilities\") pod \"community-operators-vdbjj\" (UID: \"23545a99-b853-4a78-85a6-6186dc3bf903\") " pod="openshift-marketplace/community-operators-vdbjj" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.316262 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23545a99-b853-4a78-85a6-6186dc3bf903-catalog-content\") pod \"community-operators-vdbjj\" (UID: \"23545a99-b853-4a78-85a6-6186dc3bf903\") " pod="openshift-marketplace/community-operators-vdbjj" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.316410 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.316448 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ljd6\" (UniqueName: \"kubernetes.io/projected/23545a99-b853-4a78-85a6-6186dc3bf903-kube-api-access-8ljd6\") pod \"community-operators-vdbjj\" (UID: \"23545a99-b853-4a78-85a6-6186dc3bf903\") " pod="openshift-marketplace/community-operators-vdbjj" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.316498 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ca80dd4-6294-4263-b938-188c17814556-catalog-content\") pod \"certified-operators-wbzsb\" (UID: \"1ca80dd4-6294-4263-b938-188c17814556\") " pod="openshift-marketplace/certified-operators-wbzsb" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.316570 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf74x\" (UniqueName: \"kubernetes.io/projected/1ca80dd4-6294-4263-b938-188c17814556-kube-api-access-rf74x\") pod \"certified-operators-wbzsb\" (UID: \"1ca80dd4-6294-4263-b938-188c17814556\") " pod="openshift-marketplace/certified-operators-wbzsb" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.311789 4685 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-vh7kj container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.28:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.317216 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj" podUID="926501e1-d68a-4c86-b59e-f87c43c3a9ae" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.28:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.313028 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23545a99-b853-4a78-85a6-6186dc3bf903-utilities\") pod \"community-operators-vdbjj\" (UID: \"23545a99-b853-4a78-85a6-6186dc3bf903\") " pod="openshift-marketplace/community-operators-vdbjj" Oct 13 08:47:05 crc kubenswrapper[4685]: E1013 08:47:05.317093 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:05.817071015 +0000 UTC m=+150.964946776 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.322255 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23545a99-b853-4a78-85a6-6186dc3bf903-catalog-content\") pod \"community-operators-vdbjj\" (UID: \"23545a99-b853-4a78-85a6-6186dc3bf903\") " pod="openshift-marketplace/community-operators-vdbjj" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.324812 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"d8da374427d0347590ae5343d78feb3b868534dd3dd75479e1c823c764bc21f6"} Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.324889 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"bc26cb45238b7441c19adec8885addf39ec0b198b665830bdf4fe9f6557ce0b5"} Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.349161 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f46fr" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.383419 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-w2xf4"] Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.405977 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w2xf4" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.410839 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ljd6\" (UniqueName: \"kubernetes.io/projected/23545a99-b853-4a78-85a6-6186dc3bf903-kube-api-access-8ljd6\") pod \"community-operators-vdbjj\" (UID: \"23545a99-b853-4a78-85a6-6186dc3bf903\") " pod="openshift-marketplace/community-operators-vdbjj" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.418612 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.418824 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ca80dd4-6294-4263-b938-188c17814556-utilities\") pod \"certified-operators-wbzsb\" (UID: \"1ca80dd4-6294-4263-b938-188c17814556\") " pod="openshift-marketplace/certified-operators-wbzsb" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.418894 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ca80dd4-6294-4263-b938-188c17814556-catalog-content\") pod \"certified-operators-wbzsb\" (UID: \"1ca80dd4-6294-4263-b938-188c17814556\") " pod="openshift-marketplace/certified-operators-wbzsb" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.418980 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf74x\" (UniqueName: \"kubernetes.io/projected/1ca80dd4-6294-4263-b938-188c17814556-kube-api-access-rf74x\") pod \"certified-operators-wbzsb\" (UID: \"1ca80dd4-6294-4263-b938-188c17814556\") " pod="openshift-marketplace/certified-operators-wbzsb" Oct 13 08:47:05 crc kubenswrapper[4685]: E1013 08:47:05.419339 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:05.919322161 +0000 UTC m=+151.067197922 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.419830 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ca80dd4-6294-4263-b938-188c17814556-utilities\") pod \"certified-operators-wbzsb\" (UID: \"1ca80dd4-6294-4263-b938-188c17814556\") " pod="openshift-marketplace/certified-operators-wbzsb" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.420384 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ca80dd4-6294-4263-b938-188c17814556-catalog-content\") pod \"certified-operators-wbzsb\" (UID: \"1ca80dd4-6294-4263-b938-188c17814556\") " pod="openshift-marketplace/certified-operators-wbzsb" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.467503 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf74x\" (UniqueName: \"kubernetes.io/projected/1ca80dd4-6294-4263-b938-188c17814556-kube-api-access-rf74x\") pod \"certified-operators-wbzsb\" (UID: \"1ca80dd4-6294-4263-b938-188c17814556\") " pod="openshift-marketplace/certified-operators-wbzsb" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.483443 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w2xf4"] Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.520680 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c18864ff-4530-4655-b491-25f85b5604b4-utilities\") pod \"community-operators-w2xf4\" (UID: \"c18864ff-4530-4655-b491-25f85b5604b4\") " pod="openshift-marketplace/community-operators-w2xf4" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.520718 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c18864ff-4530-4655-b491-25f85b5604b4-catalog-content\") pod \"community-operators-w2xf4\" (UID: \"c18864ff-4530-4655-b491-25f85b5604b4\") " pod="openshift-marketplace/community-operators-w2xf4" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.520968 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.521028 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4lpc\" (UniqueName: \"kubernetes.io/projected/c18864ff-4530-4655-b491-25f85b5604b4-kube-api-access-c4lpc\") pod \"community-operators-w2xf4\" (UID: \"c18864ff-4530-4655-b491-25f85b5604b4\") " pod="openshift-marketplace/community-operators-w2xf4" Oct 13 08:47:05 crc kubenswrapper[4685]: E1013 08:47:05.526364 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:06.02634502 +0000 UTC m=+151.174220971 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.526830 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wbzsb" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.559285 4685 patch_prober.go:28] interesting pod/apiserver-76f77b778f-kgrk6 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 13 08:47:05 crc kubenswrapper[4685]: [+]log ok Oct 13 08:47:05 crc kubenswrapper[4685]: [+]etcd ok Oct 13 08:47:05 crc kubenswrapper[4685]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 13 08:47:05 crc kubenswrapper[4685]: [+]poststarthook/generic-apiserver-start-informers ok Oct 13 08:47:05 crc kubenswrapper[4685]: [+]poststarthook/max-in-flight-filter ok Oct 13 08:47:05 crc kubenswrapper[4685]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 13 08:47:05 crc kubenswrapper[4685]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 13 08:47:05 crc kubenswrapper[4685]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 13 08:47:05 crc kubenswrapper[4685]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 13 08:47:05 crc kubenswrapper[4685]: [-]poststarthook/project.openshift.io-projectcache failed: reason withheld Oct 13 08:47:05 crc kubenswrapper[4685]: [-]poststarthook/project.openshift.io-projectauthorizationcache failed: reason withheld Oct 13 08:47:05 crc kubenswrapper[4685]: [+]poststarthook/openshift.io-startinformers ok Oct 13 08:47:05 crc kubenswrapper[4685]: [-]poststarthook/openshift.io-restmapperupdater failed: reason withheld Oct 13 08:47:05 crc kubenswrapper[4685]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 13 08:47:05 crc kubenswrapper[4685]: livez check failed Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.559907 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" podUID="8f247d85-48fa-49a1-9c0c-3ec7b7d202c0" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.560506 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-r96rq" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.583387 4685 patch_prober.go:28] interesting pod/router-default-5444994796-r96rq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 08:47:05 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Oct 13 08:47:05 crc kubenswrapper[4685]: [+]process-running ok Oct 13 08:47:05 crc kubenswrapper[4685]: healthz check failed Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.583465 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r96rq" podUID="0fe547bf-ea91-4546-a886-c613abeae02a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.619406 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-csdt2"] Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.620836 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-csdt2" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.622323 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.622646 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4lpc\" (UniqueName: \"kubernetes.io/projected/c18864ff-4530-4655-b491-25f85b5604b4-kube-api-access-c4lpc\") pod \"community-operators-w2xf4\" (UID: \"c18864ff-4530-4655-b491-25f85b5604b4\") " pod="openshift-marketplace/community-operators-w2xf4" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.622704 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c18864ff-4530-4655-b491-25f85b5604b4-utilities\") pod \"community-operators-w2xf4\" (UID: \"c18864ff-4530-4655-b491-25f85b5604b4\") " pod="openshift-marketplace/community-operators-w2xf4" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.622732 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c18864ff-4530-4655-b491-25f85b5604b4-catalog-content\") pod \"community-operators-w2xf4\" (UID: \"c18864ff-4530-4655-b491-25f85b5604b4\") " pod="openshift-marketplace/community-operators-w2xf4" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.623256 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c18864ff-4530-4655-b491-25f85b5604b4-catalog-content\") pod \"community-operators-w2xf4\" (UID: \"c18864ff-4530-4655-b491-25f85b5604b4\") " pod="openshift-marketplace/community-operators-w2xf4" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.623956 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c18864ff-4530-4655-b491-25f85b5604b4-utilities\") pod \"community-operators-w2xf4\" (UID: \"c18864ff-4530-4655-b491-25f85b5604b4\") " pod="openshift-marketplace/community-operators-w2xf4" Oct 13 08:47:05 crc kubenswrapper[4685]: E1013 08:47:05.624365 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:06.124336541 +0000 UTC m=+151.272212302 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.649344 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vdbjj" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.692659 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4lpc\" (UniqueName: \"kubernetes.io/projected/c18864ff-4530-4655-b491-25f85b5604b4-kube-api-access-c4lpc\") pod \"community-operators-w2xf4\" (UID: \"c18864ff-4530-4655-b491-25f85b5604b4\") " pod="openshift-marketplace/community-operators-w2xf4" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.728831 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a006d49-7467-4c6a-9024-09de1da2d8da-catalog-content\") pod \"certified-operators-csdt2\" (UID: \"2a006d49-7467-4c6a-9024-09de1da2d8da\") " pod="openshift-marketplace/certified-operators-csdt2" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.729190 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a006d49-7467-4c6a-9024-09de1da2d8da-utilities\") pod \"certified-operators-csdt2\" (UID: \"2a006d49-7467-4c6a-9024-09de1da2d8da\") " pod="openshift-marketplace/certified-operators-csdt2" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.729323 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jch9n\" (UniqueName: \"kubernetes.io/projected/2a006d49-7467-4c6a-9024-09de1da2d8da-kube-api-access-jch9n\") pod \"certified-operators-csdt2\" (UID: \"2a006d49-7467-4c6a-9024-09de1da2d8da\") " pod="openshift-marketplace/certified-operators-csdt2" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.729417 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.737994 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-csdt2"] Oct 13 08:47:05 crc kubenswrapper[4685]: E1013 08:47:05.739892 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:06.239872702 +0000 UTC m=+151.387748453 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.764440 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w2xf4" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.830709 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.831076 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jch9n\" (UniqueName: \"kubernetes.io/projected/2a006d49-7467-4c6a-9024-09de1da2d8da-kube-api-access-jch9n\") pod \"certified-operators-csdt2\" (UID: \"2a006d49-7467-4c6a-9024-09de1da2d8da\") " pod="openshift-marketplace/certified-operators-csdt2" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.831150 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a006d49-7467-4c6a-9024-09de1da2d8da-catalog-content\") pod \"certified-operators-csdt2\" (UID: \"2a006d49-7467-4c6a-9024-09de1da2d8da\") " pod="openshift-marketplace/certified-operators-csdt2" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.831206 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a006d49-7467-4c6a-9024-09de1da2d8da-utilities\") pod \"certified-operators-csdt2\" (UID: \"2a006d49-7467-4c6a-9024-09de1da2d8da\") " pod="openshift-marketplace/certified-operators-csdt2" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.831723 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a006d49-7467-4c6a-9024-09de1da2d8da-utilities\") pod \"certified-operators-csdt2\" (UID: \"2a006d49-7467-4c6a-9024-09de1da2d8da\") " pod="openshift-marketplace/certified-operators-csdt2" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.831998 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a006d49-7467-4c6a-9024-09de1da2d8da-catalog-content\") pod \"certified-operators-csdt2\" (UID: \"2a006d49-7467-4c6a-9024-09de1da2d8da\") " pod="openshift-marketplace/certified-operators-csdt2" Oct 13 08:47:05 crc kubenswrapper[4685]: E1013 08:47:05.832084 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:06.332060252 +0000 UTC m=+151.479936013 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.892620 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jch9n\" (UniqueName: \"kubernetes.io/projected/2a006d49-7467-4c6a-9024-09de1da2d8da-kube-api-access-jch9n\") pod \"certified-operators-csdt2\" (UID: \"2a006d49-7467-4c6a-9024-09de1da2d8da\") " pod="openshift-marketplace/certified-operators-csdt2" Oct 13 08:47:05 crc kubenswrapper[4685]: I1013 08:47:05.932715 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:05 crc kubenswrapper[4685]: E1013 08:47:05.933475 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:06.433451023 +0000 UTC m=+151.581326784 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.009265 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-csdt2" Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.034801 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:06 crc kubenswrapper[4685]: E1013 08:47:06.035312 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:06.535287198 +0000 UTC m=+151.683162959 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.136352 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:06 crc kubenswrapper[4685]: E1013 08:47:06.137150 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:06.637138623 +0000 UTC m=+151.785014384 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:06 crc kubenswrapper[4685]: W1013 08:47:06.200361 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-583fab93d7b9aa1b617644194a0b1db0f34523436eb7d251d1726d87abb32ad7 WatchSource:0}: Error finding container 583fab93d7b9aa1b617644194a0b1db0f34523436eb7d251d1726d87abb32ad7: Status 404 returned error can't find the container with id 583fab93d7b9aa1b617644194a0b1db0f34523436eb7d251d1726d87abb32ad7 Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.244353 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:06 crc kubenswrapper[4685]: E1013 08:47:06.244700 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:06.744685298 +0000 UTC m=+151.892561059 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.350182 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4lflr" Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.352338 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:06 crc kubenswrapper[4685]: E1013 08:47:06.353366 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:06.853354405 +0000 UTC m=+152.001230166 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.411153 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-96w58" event={"ID":"763388db-ca9e-432e-9c91-4821392af1a1","Type":"ContainerStarted","Data":"c00b97874f67022937f8ed7fa8fe824c05da691082db9e1de7e829f8d1b99a1d"} Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.414432 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"583fab93d7b9aa1b617644194a0b1db0f34523436eb7d251d1726d87abb32ad7"} Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.445168 4685 generic.go:334] "Generic (PLEG): container finished" podID="4e31f8b2-2961-4bb6-9465-9d2e66c9f000" containerID="793f9f083d5c31ff31e1c0ca7d704fdbae10d705348e730aff6a679006402e16" exitCode=0 Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.445269 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4e31f8b2-2961-4bb6-9465-9d2e66c9f000","Type":"ContainerDied","Data":"793f9f083d5c31ff31e1c0ca7d704fdbae10d705348e730aff6a679006402e16"} Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.453474 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:06 crc kubenswrapper[4685]: E1013 08:47:06.453883 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:06.953867341 +0000 UTC m=+152.101743092 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.489135 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"67f83baff0b76cbcec9f37e330e3530abf05bee0364951b33c1fe1fe4eee5c10"} Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.489191 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"2242df5c4217ce3f1b8a041a1f0a180680d52992989593d572ed90dd8c0b5673"} Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.566743 4685 patch_prober.go:28] interesting pod/router-default-5444994796-r96rq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 08:47:06 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Oct 13 08:47:06 crc kubenswrapper[4685]: [+]process-running ok Oct 13 08:47:06 crc kubenswrapper[4685]: healthz check failed Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.566859 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r96rq" podUID="0fe547bf-ea91-4546-a886-c613abeae02a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.596314 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:06 crc kubenswrapper[4685]: E1013 08:47:06.614369 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:07.114339162 +0000 UTC m=+152.262214923 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.663356 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vh7kj" Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.717371 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:06 crc kubenswrapper[4685]: E1013 08:47:06.718765 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:07.218748672 +0000 UTC m=+152.366624433 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.818964 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:06 crc kubenswrapper[4685]: E1013 08:47:06.819405 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:07.319389932 +0000 UTC m=+152.467265703 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.868620 4685 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7sx5f container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.868705 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" podUID="b8ae1ce1-bf5f-4d13-87b4-0810f63434af" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.869181 4685 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7sx5f container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.869206 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" podUID="b8ae1ce1-bf5f-4d13-87b4-0810f63434af" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Oct 13 08:47:06 crc kubenswrapper[4685]: I1013 08:47:06.919704 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:06 crc kubenswrapper[4685]: E1013 08:47:06.920198 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:07.420178956 +0000 UTC m=+152.568054717 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.021034 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:07 crc kubenswrapper[4685]: E1013 08:47:07.021818 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:07.521805994 +0000 UTC m=+152.669681755 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.106632 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wbzsb"] Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.122826 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:07 crc kubenswrapper[4685]: E1013 08:47:07.123302 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:07.623281859 +0000 UTC m=+152.771157620 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.198641 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fmmfz"] Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.199736 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fmmfz" Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.213746 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.226687 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:07 crc kubenswrapper[4685]: E1013 08:47:07.227007 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:07.726995859 +0000 UTC m=+152.874871620 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.227982 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vdbjj"] Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.261894 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fmmfz"] Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.332829 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.333708 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e74b537e-de07-4a47-85e7-e2bd5a72a11c-utilities\") pod \"redhat-marketplace-fmmfz\" (UID: \"e74b537e-de07-4a47-85e7-e2bd5a72a11c\") " pod="openshift-marketplace/redhat-marketplace-fmmfz" Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.333739 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e74b537e-de07-4a47-85e7-e2bd5a72a11c-catalog-content\") pod \"redhat-marketplace-fmmfz\" (UID: \"e74b537e-de07-4a47-85e7-e2bd5a72a11c\") " pod="openshift-marketplace/redhat-marketplace-fmmfz" Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.333778 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrcfq\" (UniqueName: \"kubernetes.io/projected/e74b537e-de07-4a47-85e7-e2bd5a72a11c-kube-api-access-rrcfq\") pod \"redhat-marketplace-fmmfz\" (UID: \"e74b537e-de07-4a47-85e7-e2bd5a72a11c\") " pod="openshift-marketplace/redhat-marketplace-fmmfz" Oct 13 08:47:07 crc kubenswrapper[4685]: E1013 08:47:07.333904 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:07.833881495 +0000 UTC m=+152.981757256 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.434950 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e74b537e-de07-4a47-85e7-e2bd5a72a11c-utilities\") pod \"redhat-marketplace-fmmfz\" (UID: \"e74b537e-de07-4a47-85e7-e2bd5a72a11c\") " pod="openshift-marketplace/redhat-marketplace-fmmfz" Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.434986 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e74b537e-de07-4a47-85e7-e2bd5a72a11c-catalog-content\") pod \"redhat-marketplace-fmmfz\" (UID: \"e74b537e-de07-4a47-85e7-e2bd5a72a11c\") " pod="openshift-marketplace/redhat-marketplace-fmmfz" Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.435015 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrcfq\" (UniqueName: \"kubernetes.io/projected/e74b537e-de07-4a47-85e7-e2bd5a72a11c-kube-api-access-rrcfq\") pod \"redhat-marketplace-fmmfz\" (UID: \"e74b537e-de07-4a47-85e7-e2bd5a72a11c\") " pod="openshift-marketplace/redhat-marketplace-fmmfz" Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.435068 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.438876 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e74b537e-de07-4a47-85e7-e2bd5a72a11c-catalog-content\") pod \"redhat-marketplace-fmmfz\" (UID: \"e74b537e-de07-4a47-85e7-e2bd5a72a11c\") " pod="openshift-marketplace/redhat-marketplace-fmmfz" Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.439237 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e74b537e-de07-4a47-85e7-e2bd5a72a11c-utilities\") pod \"redhat-marketplace-fmmfz\" (UID: \"e74b537e-de07-4a47-85e7-e2bd5a72a11c\") " pod="openshift-marketplace/redhat-marketplace-fmmfz" Oct 13 08:47:07 crc kubenswrapper[4685]: E1013 08:47:07.440025 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:07.940005476 +0000 UTC m=+153.087881237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.498551 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrcfq\" (UniqueName: \"kubernetes.io/projected/e74b537e-de07-4a47-85e7-e2bd5a72a11c-kube-api-access-rrcfq\") pod \"redhat-marketplace-fmmfz\" (UID: \"e74b537e-de07-4a47-85e7-e2bd5a72a11c\") " pod="openshift-marketplace/redhat-marketplace-fmmfz" Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.521656 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbzsb" event={"ID":"1ca80dd4-6294-4263-b938-188c17814556","Type":"ContainerStarted","Data":"1f02035b8ba6c426061e6d3734c23df93ad75ccd361dae247d47f4f28f0c79a6"} Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.531509 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"511c2612081f5c353f98f61be5fdf93f82f2f59644ec9a97f0d9a79982c72b19"} Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.533082 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.536729 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:07 crc kubenswrapper[4685]: E1013 08:47:07.537164 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:08.037145912 +0000 UTC m=+153.185021673 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.537184 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w2xf4"] Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.550777 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fmmfz" Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.566106 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-96w58" event={"ID":"763388db-ca9e-432e-9c91-4821392af1a1","Type":"ContainerStarted","Data":"de7e3638bde01087bf2906bcdcf59a6746f16312811ae1fa67b0c9c641fb26f6"} Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.566185 4685 patch_prober.go:28] interesting pod/router-default-5444994796-r96rq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 08:47:07 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Oct 13 08:47:07 crc kubenswrapper[4685]: [+]process-running ok Oct 13 08:47:07 crc kubenswrapper[4685]: healthz check failed Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.566213 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r96rq" podUID="0fe547bf-ea91-4546-a886-c613abeae02a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.593734 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdbjj" event={"ID":"23545a99-b853-4a78-85a6-6186dc3bf903","Type":"ContainerStarted","Data":"c36885e110b59fb5a1bc436423fc8ca80824ee5b58e3d39ebb675a514d2496f9"} Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.607982 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wmkhb"] Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.609303 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wmkhb" Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.638019 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:07 crc kubenswrapper[4685]: E1013 08:47:07.639429 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:08.13940553 +0000 UTC m=+153.287281461 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.666051 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wmkhb"] Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.739817 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:07 crc kubenswrapper[4685]: E1013 08:47:07.740148 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:08.240105581 +0000 UTC m=+153.387981342 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.740574 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btnt9\" (UniqueName: \"kubernetes.io/projected/8beb316d-efe3-427b-a377-c84063b2b5d6-kube-api-access-btnt9\") pod \"redhat-marketplace-wmkhb\" (UID: \"8beb316d-efe3-427b-a377-c84063b2b5d6\") " pod="openshift-marketplace/redhat-marketplace-wmkhb" Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.740611 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8beb316d-efe3-427b-a377-c84063b2b5d6-catalog-content\") pod \"redhat-marketplace-wmkhb\" (UID: \"8beb316d-efe3-427b-a377-c84063b2b5d6\") " pod="openshift-marketplace/redhat-marketplace-wmkhb" Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.740653 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8beb316d-efe3-427b-a377-c84063b2b5d6-utilities\") pod \"redhat-marketplace-wmkhb\" (UID: \"8beb316d-efe3-427b-a377-c84063b2b5d6\") " pod="openshift-marketplace/redhat-marketplace-wmkhb" Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.740714 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:07 crc kubenswrapper[4685]: E1013 08:47:07.741192 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:08.241169002 +0000 UTC m=+153.389044763 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.752603 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-csdt2"] Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.843332 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.843587 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btnt9\" (UniqueName: \"kubernetes.io/projected/8beb316d-efe3-427b-a377-c84063b2b5d6-kube-api-access-btnt9\") pod \"redhat-marketplace-wmkhb\" (UID: \"8beb316d-efe3-427b-a377-c84063b2b5d6\") " pod="openshift-marketplace/redhat-marketplace-wmkhb" Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.843614 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8beb316d-efe3-427b-a377-c84063b2b5d6-catalog-content\") pod \"redhat-marketplace-wmkhb\" (UID: \"8beb316d-efe3-427b-a377-c84063b2b5d6\") " pod="openshift-marketplace/redhat-marketplace-wmkhb" Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.843660 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8beb316d-efe3-427b-a377-c84063b2b5d6-utilities\") pod \"redhat-marketplace-wmkhb\" (UID: \"8beb316d-efe3-427b-a377-c84063b2b5d6\") " pod="openshift-marketplace/redhat-marketplace-wmkhb" Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.844082 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8beb316d-efe3-427b-a377-c84063b2b5d6-utilities\") pod \"redhat-marketplace-wmkhb\" (UID: \"8beb316d-efe3-427b-a377-c84063b2b5d6\") " pod="openshift-marketplace/redhat-marketplace-wmkhb" Oct 13 08:47:07 crc kubenswrapper[4685]: E1013 08:47:07.844280 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:08.344265515 +0000 UTC m=+153.492141276 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.844764 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8beb316d-efe3-427b-a377-c84063b2b5d6-catalog-content\") pod \"redhat-marketplace-wmkhb\" (UID: \"8beb316d-efe3-427b-a377-c84063b2b5d6\") " pod="openshift-marketplace/redhat-marketplace-wmkhb" Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.876304 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btnt9\" (UniqueName: \"kubernetes.io/projected/8beb316d-efe3-427b-a377-c84063b2b5d6-kube-api-access-btnt9\") pod \"redhat-marketplace-wmkhb\" (UID: \"8beb316d-efe3-427b-a377-c84063b2b5d6\") " pod="openshift-marketplace/redhat-marketplace-wmkhb" Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.949512 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:07 crc kubenswrapper[4685]: E1013 08:47:07.949873 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:08.449860071 +0000 UTC m=+153.597735832 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:07 crc kubenswrapper[4685]: I1013 08:47:07.984797 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wmkhb" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.057762 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:08 crc kubenswrapper[4685]: E1013 08:47:08.058347 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:08.558324092 +0000 UTC m=+153.706199853 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.159713 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:08 crc kubenswrapper[4685]: E1013 08:47:08.160181 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:08.660168288 +0000 UTC m=+153.808044049 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.192018 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-v6zv9"] Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.193387 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v6zv9" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.198446 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.228554 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v6zv9"] Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.262520 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.262892 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drl5f\" (UniqueName: \"kubernetes.io/projected/75521184-74d3-44b9-a410-b0b0fafa4384-kube-api-access-drl5f\") pod \"redhat-operators-v6zv9\" (UID: \"75521184-74d3-44b9-a410-b0b0fafa4384\") " pod="openshift-marketplace/redhat-operators-v6zv9" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.262960 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75521184-74d3-44b9-a410-b0b0fafa4384-utilities\") pod \"redhat-operators-v6zv9\" (UID: \"75521184-74d3-44b9-a410-b0b0fafa4384\") " pod="openshift-marketplace/redhat-operators-v6zv9" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.262991 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75521184-74d3-44b9-a410-b0b0fafa4384-catalog-content\") pod \"redhat-operators-v6zv9\" (UID: \"75521184-74d3-44b9-a410-b0b0fafa4384\") " pod="openshift-marketplace/redhat-operators-v6zv9" Oct 13 08:47:08 crc kubenswrapper[4685]: E1013 08:47:08.263110 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:08.763095204 +0000 UTC m=+153.910970955 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.368099 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75521184-74d3-44b9-a410-b0b0fafa4384-utilities\") pod \"redhat-operators-v6zv9\" (UID: \"75521184-74d3-44b9-a410-b0b0fafa4384\") " pod="openshift-marketplace/redhat-operators-v6zv9" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.368183 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75521184-74d3-44b9-a410-b0b0fafa4384-catalog-content\") pod \"redhat-operators-v6zv9\" (UID: \"75521184-74d3-44b9-a410-b0b0fafa4384\") " pod="openshift-marketplace/redhat-operators-v6zv9" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.368289 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.368340 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drl5f\" (UniqueName: \"kubernetes.io/projected/75521184-74d3-44b9-a410-b0b0fafa4384-kube-api-access-drl5f\") pod \"redhat-operators-v6zv9\" (UID: \"75521184-74d3-44b9-a410-b0b0fafa4384\") " pod="openshift-marketplace/redhat-operators-v6zv9" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.368690 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75521184-74d3-44b9-a410-b0b0fafa4384-utilities\") pod \"redhat-operators-v6zv9\" (UID: \"75521184-74d3-44b9-a410-b0b0fafa4384\") " pod="openshift-marketplace/redhat-operators-v6zv9" Oct 13 08:47:08 crc kubenswrapper[4685]: E1013 08:47:08.369109 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:08.869084112 +0000 UTC m=+154.016960043 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.369846 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75521184-74d3-44b9-a410-b0b0fafa4384-catalog-content\") pod \"redhat-operators-v6zv9\" (UID: \"75521184-74d3-44b9-a410-b0b0fafa4384\") " pod="openshift-marketplace/redhat-operators-v6zv9" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.397287 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.408430 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drl5f\" (UniqueName: \"kubernetes.io/projected/75521184-74d3-44b9-a410-b0b0fafa4384-kube-api-access-drl5f\") pod \"redhat-operators-v6zv9\" (UID: \"75521184-74d3-44b9-a410-b0b0fafa4384\") " pod="openshift-marketplace/redhat-operators-v6zv9" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.469119 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4e31f8b2-2961-4bb6-9465-9d2e66c9f000-kubelet-dir\") pod \"4e31f8b2-2961-4bb6-9465-9d2e66c9f000\" (UID: \"4e31f8b2-2961-4bb6-9465-9d2e66c9f000\") " Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.469392 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.469448 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4e31f8b2-2961-4bb6-9465-9d2e66c9f000-kube-api-access\") pod \"4e31f8b2-2961-4bb6-9465-9d2e66c9f000\" (UID: \"4e31f8b2-2961-4bb6-9465-9d2e66c9f000\") " Oct 13 08:47:08 crc kubenswrapper[4685]: E1013 08:47:08.470184 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:08.970134854 +0000 UTC m=+154.118010615 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.470249 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4e31f8b2-2961-4bb6-9465-9d2e66c9f000-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4e31f8b2-2961-4bb6-9465-9d2e66c9f000" (UID: "4e31f8b2-2961-4bb6-9465-9d2e66c9f000"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.483795 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fmmfz"] Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.484235 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e31f8b2-2961-4bb6-9465-9d2e66c9f000-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4e31f8b2-2961-4bb6-9465-9d2e66c9f000" (UID: "4e31f8b2-2961-4bb6-9465-9d2e66c9f000"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:47:08 crc kubenswrapper[4685]: W1013 08:47:08.523849 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode74b537e_de07_4a47_85e7_e2bd5a72a11c.slice/crio-75a9741253f9702f47ad595be38cbfb122c4db6ec6b6ea1ec0077314f161a354 WatchSource:0}: Error finding container 75a9741253f9702f47ad595be38cbfb122c4db6ec6b6ea1ec0077314f161a354: Status 404 returned error can't find the container with id 75a9741253f9702f47ad595be38cbfb122c4db6ec6b6ea1ec0077314f161a354 Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.573259 4685 patch_prober.go:28] interesting pod/router-default-5444994796-r96rq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 08:47:08 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Oct 13 08:47:08 crc kubenswrapper[4685]: [+]process-running ok Oct 13 08:47:08 crc kubenswrapper[4685]: healthz check failed Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.573314 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r96rq" podUID="0fe547bf-ea91-4546-a886-c613abeae02a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.577057 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bnb5g"] Oct 13 08:47:08 crc kubenswrapper[4685]: E1013 08:47:08.577269 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e31f8b2-2961-4bb6-9465-9d2e66c9f000" containerName="pruner" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.577279 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e31f8b2-2961-4bb6-9465-9d2e66c9f000" containerName="pruner" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.577369 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e31f8b2-2961-4bb6-9465-9d2e66c9f000" containerName="pruner" Oct 13 08:47:08 crc kubenswrapper[4685]: E1013 08:47:08.578281 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:09.078264166 +0000 UTC m=+154.226139927 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.582167 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bnb5g" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.582689 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.582854 4685 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4e31f8b2-2961-4bb6-9465-9d2e66c9f000-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.582869 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4e31f8b2-2961-4bb6-9465-9d2e66c9f000-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.616049 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bnb5g"] Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.620404 4685 generic.go:334] "Generic (PLEG): container finished" podID="c18864ff-4530-4655-b491-25f85b5604b4" containerID="b9f3050114c1ef462b6fcab1e3804bef5187bb91acf8d966c210310f3232a4fa" exitCode=0 Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.620482 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2xf4" event={"ID":"c18864ff-4530-4655-b491-25f85b5604b4","Type":"ContainerDied","Data":"b9f3050114c1ef462b6fcab1e3804bef5187bb91acf8d966c210310f3232a4fa"} Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.620514 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2xf4" event={"ID":"c18864ff-4530-4655-b491-25f85b5604b4","Type":"ContainerStarted","Data":"d16dc906b086440cea6b3ceb1f6f831bd740160c3f79845aa559fbc6d7efa573"} Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.623044 4685 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.632336 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdbjj" event={"ID":"23545a99-b853-4a78-85a6-6186dc3bf903","Type":"ContainerDied","Data":"b4fd611ac286faa6b033f5fbf108e793a8af2ebf53102ce520ec2303d8cddc36"} Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.634988 4685 generic.go:334] "Generic (PLEG): container finished" podID="23545a99-b853-4a78-85a6-6186dc3bf903" containerID="b4fd611ac286faa6b033f5fbf108e793a8af2ebf53102ce520ec2303d8cddc36" exitCode=0 Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.644147 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4e31f8b2-2961-4bb6-9465-9d2e66c9f000","Type":"ContainerDied","Data":"ec388fedf392452f1b998dee1ad0b506559cb41e8f7539c1ec383c05089734f2"} Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.644186 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec388fedf392452f1b998dee1ad0b506559cb41e8f7539c1ec383c05089734f2" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.644261 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.648574 4685 generic.go:334] "Generic (PLEG): container finished" podID="2a006d49-7467-4c6a-9024-09de1da2d8da" containerID="996e1c2258995f0e196833ae4d02fed1511f882f91438ac534fe16b223f11b14" exitCode=0 Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.648653 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csdt2" event={"ID":"2a006d49-7467-4c6a-9024-09de1da2d8da","Type":"ContainerDied","Data":"996e1c2258995f0e196833ae4d02fed1511f882f91438ac534fe16b223f11b14"} Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.648683 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csdt2" event={"ID":"2a006d49-7467-4c6a-9024-09de1da2d8da","Type":"ContainerStarted","Data":"46a94fdcf5d71b5172f76dd666d6bca2896e9c295a66888988381093d17fee2e"} Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.649317 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fmmfz" event={"ID":"e74b537e-de07-4a47-85e7-e2bd5a72a11c","Type":"ContainerStarted","Data":"75a9741253f9702f47ad595be38cbfb122c4db6ec6b6ea1ec0077314f161a354"} Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.650563 4685 generic.go:334] "Generic (PLEG): container finished" podID="3672a087-9d9b-45ee-8f80-686959b395d6" containerID="c41f5b326a5ed917fc98da2be46741a8baa16dd62bb2a2fb843af8060340137f" exitCode=0 Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.650617 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn" event={"ID":"3672a087-9d9b-45ee-8f80-686959b395d6","Type":"ContainerDied","Data":"c41f5b326a5ed917fc98da2be46741a8baa16dd62bb2a2fb843af8060340137f"} Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.651683 4685 generic.go:334] "Generic (PLEG): container finished" podID="1ca80dd4-6294-4263-b938-188c17814556" containerID="a355a3c958302c1e868888905c2e735fa87447330184e854d0c34cbd9627f830" exitCode=0 Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.651734 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbzsb" event={"ID":"1ca80dd4-6294-4263-b938-188c17814556","Type":"ContainerDied","Data":"a355a3c958302c1e868888905c2e735fa87447330184e854d0c34cbd9627f830"} Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.657031 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v6zv9" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.692335 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.692618 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a9fb57f-9b71-4922-aa37-2661e199871d-utilities\") pod \"redhat-operators-bnb5g\" (UID: \"9a9fb57f-9b71-4922-aa37-2661e199871d\") " pod="openshift-marketplace/redhat-operators-bnb5g" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.692675 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a9fb57f-9b71-4922-aa37-2661e199871d-catalog-content\") pod \"redhat-operators-bnb5g\" (UID: \"9a9fb57f-9b71-4922-aa37-2661e199871d\") " pod="openshift-marketplace/redhat-operators-bnb5g" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.692702 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tjlv\" (UniqueName: \"kubernetes.io/projected/9a9fb57f-9b71-4922-aa37-2661e199871d-kube-api-access-2tjlv\") pod \"redhat-operators-bnb5g\" (UID: \"9a9fb57f-9b71-4922-aa37-2661e199871d\") " pod="openshift-marketplace/redhat-operators-bnb5g" Oct 13 08:47:08 crc kubenswrapper[4685]: E1013 08:47:08.692842 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:09.192826697 +0000 UTC m=+154.340702458 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.701843 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wmkhb"] Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.793841 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a9fb57f-9b71-4922-aa37-2661e199871d-catalog-content\") pod \"redhat-operators-bnb5g\" (UID: \"9a9fb57f-9b71-4922-aa37-2661e199871d\") " pod="openshift-marketplace/redhat-operators-bnb5g" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.793905 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tjlv\" (UniqueName: \"kubernetes.io/projected/9a9fb57f-9b71-4922-aa37-2661e199871d-kube-api-access-2tjlv\") pod \"redhat-operators-bnb5g\" (UID: \"9a9fb57f-9b71-4922-aa37-2661e199871d\") " pod="openshift-marketplace/redhat-operators-bnb5g" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.794013 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a9fb57f-9b71-4922-aa37-2661e199871d-utilities\") pod \"redhat-operators-bnb5g\" (UID: \"9a9fb57f-9b71-4922-aa37-2661e199871d\") " pod="openshift-marketplace/redhat-operators-bnb5g" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.794065 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.796411 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a9fb57f-9b71-4922-aa37-2661e199871d-catalog-content\") pod \"redhat-operators-bnb5g\" (UID: \"9a9fb57f-9b71-4922-aa37-2661e199871d\") " pod="openshift-marketplace/redhat-operators-bnb5g" Oct 13 08:47:08 crc kubenswrapper[4685]: E1013 08:47:08.796698 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:09.296684442 +0000 UTC m=+154.444560203 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.797092 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a9fb57f-9b71-4922-aa37-2661e199871d-utilities\") pod \"redhat-operators-bnb5g\" (UID: \"9a9fb57f-9b71-4922-aa37-2661e199871d\") " pod="openshift-marketplace/redhat-operators-bnb5g" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.822081 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tjlv\" (UniqueName: \"kubernetes.io/projected/9a9fb57f-9b71-4922-aa37-2661e199871d-kube-api-access-2tjlv\") pod \"redhat-operators-bnb5g\" (UID: \"9a9fb57f-9b71-4922-aa37-2661e199871d\") " pod="openshift-marketplace/redhat-operators-bnb5g" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.860948 4685 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.895470 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:08 crc kubenswrapper[4685]: E1013 08:47:08.895844 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:09.395811887 +0000 UTC m=+154.543687658 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.896161 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:08 crc kubenswrapper[4685]: E1013 08:47:08.896684 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:09.396673113 +0000 UTC m=+154.544548874 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.997894 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:08 crc kubenswrapper[4685]: E1013 08:47:08.998160 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:09.498132036 +0000 UTC m=+154.646007787 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:08 crc kubenswrapper[4685]: I1013 08:47:08.998241 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:08 crc kubenswrapper[4685]: E1013 08:47:08.998692 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:09.498678682 +0000 UTC m=+154.646554443 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.003834 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bnb5g" Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.065801 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v6zv9"] Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.099654 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:09 crc kubenswrapper[4685]: E1013 08:47:09.099936 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:09.599905459 +0000 UTC m=+154.747781220 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.099964 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:09 crc kubenswrapper[4685]: E1013 08:47:09.100249 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:09.600240379 +0000 UTC m=+154.748116140 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.201378 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:09 crc kubenswrapper[4685]: E1013 08:47:09.202456 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:09.702427695 +0000 UTC m=+154.850303456 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.304419 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:09 crc kubenswrapper[4685]: E1013 08:47:09.305401 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:09.805387552 +0000 UTC m=+154.953263313 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.320924 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bnb5g"] Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.405482 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:09 crc kubenswrapper[4685]: E1013 08:47:09.406040 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:09.906020572 +0000 UTC m=+155.053896333 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.507142 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:09 crc kubenswrapper[4685]: E1013 08:47:09.508392 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:10.008375612 +0000 UTC m=+155.156251373 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.556154 4685 patch_prober.go:28] interesting pod/router-default-5444994796-r96rq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 08:47:09 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Oct 13 08:47:09 crc kubenswrapper[4685]: [+]process-running ok Oct 13 08:47:09 crc kubenswrapper[4685]: healthz check failed Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.556225 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r96rq" podUID="0fe547bf-ea91-4546-a886-c613abeae02a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.609687 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:09 crc kubenswrapper[4685]: E1013 08:47:09.609996 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:10.109946469 +0000 UTC m=+155.257822230 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.610277 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:09 crc kubenswrapper[4685]: E1013 08:47:09.610753 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:10.110732822 +0000 UTC m=+155.258608583 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.712028 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:09 crc kubenswrapper[4685]: E1013 08:47:09.712313 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 08:47:10.212277439 +0000 UTC m=+155.360153210 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.712394 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:09 crc kubenswrapper[4685]: E1013 08:47:09.712885 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 08:47:10.212867087 +0000 UTC m=+155.360742838 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8jzsj" (UID: "70fe2241-a755-4556-a84a-733306f90d66") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.741144 4685 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-13T08:47:08.860982256Z","Handler":null,"Name":""} Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.758729 4685 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.758790 4685 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.773275 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.797827 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-kgrk6" Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.813525 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.822188 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.879371 4685 generic.go:334] "Generic (PLEG): container finished" podID="8beb316d-efe3-427b-a377-c84063b2b5d6" containerID="085a1258cf27bb562aa13a64bd7c5df03a897160a7f63f1e37e21487aef8d60e" exitCode=0 Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.879458 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmkhb" event={"ID":"8beb316d-efe3-427b-a377-c84063b2b5d6","Type":"ContainerDied","Data":"085a1258cf27bb562aa13a64bd7c5df03a897160a7f63f1e37e21487aef8d60e"} Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.879506 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmkhb" event={"ID":"8beb316d-efe3-427b-a377-c84063b2b5d6","Type":"ContainerStarted","Data":"034a890660d9f39279f8b27013e32a015e13bce374b32fec9d733d5db39758d2"} Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.913636 4685 generic.go:334] "Generic (PLEG): container finished" podID="9a9fb57f-9b71-4922-aa37-2661e199871d" containerID="c056ab57ddd7b0cf23825e2a8a185e5c971a0ec89eac9b83db07e0032008a567" exitCode=0 Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.913775 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnb5g" event={"ID":"9a9fb57f-9b71-4922-aa37-2661e199871d","Type":"ContainerDied","Data":"c056ab57ddd7b0cf23825e2a8a185e5c971a0ec89eac9b83db07e0032008a567"} Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.913816 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnb5g" event={"ID":"9a9fb57f-9b71-4922-aa37-2661e199871d","Type":"ContainerStarted","Data":"9cdf134f5276561e6ec5d70e073b1b8bf61059d01d473173797f4897637c0ea6"} Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.923552 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.936803 4685 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.936860 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.939178 4685 generic.go:334] "Generic (PLEG): container finished" podID="e74b537e-de07-4a47-85e7-e2bd5a72a11c" containerID="87412e9e92501e922f8d5586ec8004940e405e413e2494f1859a71fc3d072cce" exitCode=0 Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.939293 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fmmfz" event={"ID":"e74b537e-de07-4a47-85e7-e2bd5a72a11c","Type":"ContainerDied","Data":"87412e9e92501e922f8d5586ec8004940e405e413e2494f1859a71fc3d072cce"} Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.967928 4685 generic.go:334] "Generic (PLEG): container finished" podID="75521184-74d3-44b9-a410-b0b0fafa4384" containerID="71df0cf490c8514d2a1280ace7996d87626606dbfcc11bea3502cc3fc2386186" exitCode=0 Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.968025 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6zv9" event={"ID":"75521184-74d3-44b9-a410-b0b0fafa4384","Type":"ContainerDied","Data":"71df0cf490c8514d2a1280ace7996d87626606dbfcc11bea3502cc3fc2386186"} Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.968071 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6zv9" event={"ID":"75521184-74d3-44b9-a410-b0b0fafa4384","Type":"ContainerStarted","Data":"f53264f68a3d9bc80305a3b0bf45176413abd2e224dccf6c8f282f9b6ed05522"} Oct 13 08:47:09 crc kubenswrapper[4685]: I1013 08:47:09.997293 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-96w58" event={"ID":"763388db-ca9e-432e-9c91-4821392af1a1","Type":"ContainerStarted","Data":"948498248d152f1d27af273e4dd9d0edadba5622c5f48efc28db38ef1b0cb0c1"} Oct 13 08:47:10 crc kubenswrapper[4685]: I1013 08:47:10.117402 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-96w58" podStartSLOduration=17.117381462 podStartE2EDuration="17.117381462s" podCreationTimestamp="2025-10-13 08:46:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:10.111615752 +0000 UTC m=+155.259491513" watchObservedRunningTime="2025-10-13 08:47:10.117381462 +0000 UTC m=+155.265257223" Oct 13 08:47:10 crc kubenswrapper[4685]: I1013 08:47:10.327576 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8jzsj\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:10 crc kubenswrapper[4685]: I1013 08:47:10.345091 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:10 crc kubenswrapper[4685]: I1013 08:47:10.557134 4685 patch_prober.go:28] interesting pod/router-default-5444994796-r96rq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 08:47:10 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Oct 13 08:47:10 crc kubenswrapper[4685]: [+]process-running ok Oct 13 08:47:10 crc kubenswrapper[4685]: healthz check failed Oct 13 08:47:10 crc kubenswrapper[4685]: I1013 08:47:10.557595 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r96rq" podUID="0fe547bf-ea91-4546-a886-c613abeae02a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 08:47:10 crc kubenswrapper[4685]: I1013 08:47:10.570586 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn" Oct 13 08:47:10 crc kubenswrapper[4685]: I1013 08:47:10.634977 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrpk2\" (UniqueName: \"kubernetes.io/projected/3672a087-9d9b-45ee-8f80-686959b395d6-kube-api-access-xrpk2\") pod \"3672a087-9d9b-45ee-8f80-686959b395d6\" (UID: \"3672a087-9d9b-45ee-8f80-686959b395d6\") " Oct 13 08:47:10 crc kubenswrapper[4685]: I1013 08:47:10.635065 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3672a087-9d9b-45ee-8f80-686959b395d6-secret-volume\") pod \"3672a087-9d9b-45ee-8f80-686959b395d6\" (UID: \"3672a087-9d9b-45ee-8f80-686959b395d6\") " Oct 13 08:47:10 crc kubenswrapper[4685]: I1013 08:47:10.635142 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3672a087-9d9b-45ee-8f80-686959b395d6-config-volume\") pod \"3672a087-9d9b-45ee-8f80-686959b395d6\" (UID: \"3672a087-9d9b-45ee-8f80-686959b395d6\") " Oct 13 08:47:10 crc kubenswrapper[4685]: I1013 08:47:10.636895 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3672a087-9d9b-45ee-8f80-686959b395d6-config-volume" (OuterVolumeSpecName: "config-volume") pod "3672a087-9d9b-45ee-8f80-686959b395d6" (UID: "3672a087-9d9b-45ee-8f80-686959b395d6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:47:10 crc kubenswrapper[4685]: I1013 08:47:10.669798 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3672a087-9d9b-45ee-8f80-686959b395d6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3672a087-9d9b-45ee-8f80-686959b395d6" (UID: "3672a087-9d9b-45ee-8f80-686959b395d6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:47:10 crc kubenswrapper[4685]: I1013 08:47:10.670617 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3672a087-9d9b-45ee-8f80-686959b395d6-kube-api-access-xrpk2" (OuterVolumeSpecName: "kube-api-access-xrpk2") pod "3672a087-9d9b-45ee-8f80-686959b395d6" (UID: "3672a087-9d9b-45ee-8f80-686959b395d6"). InnerVolumeSpecName "kube-api-access-xrpk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:47:10 crc kubenswrapper[4685]: I1013 08:47:10.736609 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrpk2\" (UniqueName: \"kubernetes.io/projected/3672a087-9d9b-45ee-8f80-686959b395d6-kube-api-access-xrpk2\") on node \"crc\" DevicePath \"\"" Oct 13 08:47:10 crc kubenswrapper[4685]: I1013 08:47:10.736701 4685 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3672a087-9d9b-45ee-8f80-686959b395d6-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 13 08:47:10 crc kubenswrapper[4685]: I1013 08:47:10.736712 4685 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3672a087-9d9b-45ee-8f80-686959b395d6-config-volume\") on node \"crc\" DevicePath \"\"" Oct 13 08:47:10 crc kubenswrapper[4685]: I1013 08:47:10.857891 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8jzsj"] Oct 13 08:47:11 crc kubenswrapper[4685]: I1013 08:47:11.117450 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" event={"ID":"70fe2241-a755-4556-a84a-733306f90d66","Type":"ContainerStarted","Data":"51c99adb4fc23a710abd311a3cb9461ff320c80e96227cc77e7fbd53d893481e"} Oct 13 08:47:11 crc kubenswrapper[4685]: I1013 08:47:11.150988 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn" Oct 13 08:47:11 crc kubenswrapper[4685]: I1013 08:47:11.151727 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn" event={"ID":"3672a087-9d9b-45ee-8f80-686959b395d6","Type":"ContainerDied","Data":"36e26e5c5a3fa5ed680abafb9f41cc8050b80f6b5327664df8fa08c6b89d51a6"} Oct 13 08:47:11 crc kubenswrapper[4685]: I1013 08:47:11.152596 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36e26e5c5a3fa5ed680abafb9f41cc8050b80f6b5327664df8fa08c6b89d51a6" Oct 13 08:47:11 crc kubenswrapper[4685]: I1013 08:47:11.543773 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 13 08:47:11 crc kubenswrapper[4685]: I1013 08:47:11.555375 4685 patch_prober.go:28] interesting pod/router-default-5444994796-r96rq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 08:47:11 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Oct 13 08:47:11 crc kubenswrapper[4685]: [+]process-running ok Oct 13 08:47:11 crc kubenswrapper[4685]: healthz check failed Oct 13 08:47:11 crc kubenswrapper[4685]: I1013 08:47:11.555452 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r96rq" podUID="0fe547bf-ea91-4546-a886-c613abeae02a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 08:47:11 crc kubenswrapper[4685]: I1013 08:47:11.629001 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-cnh25" Oct 13 08:47:12 crc kubenswrapper[4685]: I1013 08:47:12.219168 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" event={"ID":"70fe2241-a755-4556-a84a-733306f90d66","Type":"ContainerStarted","Data":"b5b59afa9d3bc7339ac45f529079822e5e870a3bf5241195fdb3709b89683243"} Oct 13 08:47:12 crc kubenswrapper[4685]: I1013 08:47:12.220423 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:12 crc kubenswrapper[4685]: I1013 08:47:12.248597 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" podStartSLOduration=136.248575589 podStartE2EDuration="2m16.248575589s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:12.244285171 +0000 UTC m=+157.392160932" watchObservedRunningTime="2025-10-13 08:47:12.248575589 +0000 UTC m=+157.396451350" Oct 13 08:47:12 crc kubenswrapper[4685]: I1013 08:47:12.557660 4685 patch_prober.go:28] interesting pod/router-default-5444994796-r96rq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 08:47:12 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Oct 13 08:47:12 crc kubenswrapper[4685]: [+]process-running ok Oct 13 08:47:12 crc kubenswrapper[4685]: healthz check failed Oct 13 08:47:12 crc kubenswrapper[4685]: I1013 08:47:12.557722 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r96rq" podUID="0fe547bf-ea91-4546-a886-c613abeae02a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 08:47:13 crc kubenswrapper[4685]: I1013 08:47:13.092228 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 13 08:47:13 crc kubenswrapper[4685]: E1013 08:47:13.092512 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3672a087-9d9b-45ee-8f80-686959b395d6" containerName="collect-profiles" Oct 13 08:47:13 crc kubenswrapper[4685]: I1013 08:47:13.092527 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="3672a087-9d9b-45ee-8f80-686959b395d6" containerName="collect-profiles" Oct 13 08:47:13 crc kubenswrapper[4685]: I1013 08:47:13.092632 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="3672a087-9d9b-45ee-8f80-686959b395d6" containerName="collect-profiles" Oct 13 08:47:13 crc kubenswrapper[4685]: I1013 08:47:13.093128 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 13 08:47:13 crc kubenswrapper[4685]: I1013 08:47:13.103882 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 13 08:47:13 crc kubenswrapper[4685]: I1013 08:47:13.103953 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 13 08:47:13 crc kubenswrapper[4685]: I1013 08:47:13.108785 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 13 08:47:13 crc kubenswrapper[4685]: I1013 08:47:13.227725 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/acf8bfed-51a6-4317-a60e-8b8ff193f47d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"acf8bfed-51a6-4317-a60e-8b8ff193f47d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 13 08:47:13 crc kubenswrapper[4685]: I1013 08:47:13.227786 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/acf8bfed-51a6-4317-a60e-8b8ff193f47d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"acf8bfed-51a6-4317-a60e-8b8ff193f47d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 13 08:47:13 crc kubenswrapper[4685]: I1013 08:47:13.329657 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/acf8bfed-51a6-4317-a60e-8b8ff193f47d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"acf8bfed-51a6-4317-a60e-8b8ff193f47d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 13 08:47:13 crc kubenswrapper[4685]: I1013 08:47:13.329795 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/acf8bfed-51a6-4317-a60e-8b8ff193f47d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"acf8bfed-51a6-4317-a60e-8b8ff193f47d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 13 08:47:13 crc kubenswrapper[4685]: I1013 08:47:13.330056 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/acf8bfed-51a6-4317-a60e-8b8ff193f47d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"acf8bfed-51a6-4317-a60e-8b8ff193f47d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 13 08:47:13 crc kubenswrapper[4685]: I1013 08:47:13.380870 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/acf8bfed-51a6-4317-a60e-8b8ff193f47d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"acf8bfed-51a6-4317-a60e-8b8ff193f47d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 13 08:47:13 crc kubenswrapper[4685]: I1013 08:47:13.435283 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 13 08:47:13 crc kubenswrapper[4685]: I1013 08:47:13.556282 4685 patch_prober.go:28] interesting pod/router-default-5444994796-r96rq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 08:47:13 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Oct 13 08:47:13 crc kubenswrapper[4685]: [+]process-running ok Oct 13 08:47:13 crc kubenswrapper[4685]: healthz check failed Oct 13 08:47:13 crc kubenswrapper[4685]: I1013 08:47:13.556395 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r96rq" podUID="0fe547bf-ea91-4546-a886-c613abeae02a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 08:47:13 crc kubenswrapper[4685]: I1013 08:47:13.986752 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 13 08:47:14 crc kubenswrapper[4685]: I1013 08:47:14.256893 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"acf8bfed-51a6-4317-a60e-8b8ff193f47d","Type":"ContainerStarted","Data":"f2387814581fc423d7d6995f3b82f628e8e60ef7025ab894f2d3a1f0f558debc"} Oct 13 08:47:14 crc kubenswrapper[4685]: I1013 08:47:14.556994 4685 patch_prober.go:28] interesting pod/router-default-5444994796-r96rq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 08:47:14 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Oct 13 08:47:14 crc kubenswrapper[4685]: [+]process-running ok Oct 13 08:47:14 crc kubenswrapper[4685]: healthz check failed Oct 13 08:47:14 crc kubenswrapper[4685]: I1013 08:47:14.557046 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r96rq" podUID="0fe547bf-ea91-4546-a886-c613abeae02a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 08:47:14 crc kubenswrapper[4685]: I1013 08:47:14.938132 4685 patch_prober.go:28] interesting pod/console-f9d7485db-gx6qj container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Oct 13 08:47:14 crc kubenswrapper[4685]: I1013 08:47:14.938187 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-gx6qj" podUID="16484304-6976-45e8-a495-f1d2ee367f0f" containerName="console" probeResult="failure" output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" Oct 13 08:47:15 crc kubenswrapper[4685]: I1013 08:47:15.248784 4685 patch_prober.go:28] interesting pod/downloads-7954f5f757-2kbsj container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 13 08:47:15 crc kubenswrapper[4685]: I1013 08:47:15.250073 4685 patch_prober.go:28] interesting pod/downloads-7954f5f757-2kbsj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 13 08:47:15 crc kubenswrapper[4685]: I1013 08:47:15.250124 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2kbsj" podUID="44a7f769-9a31-4555-be16-51e310ac34e3" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 13 08:47:15 crc kubenswrapper[4685]: I1013 08:47:15.248891 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-2kbsj" podUID="44a7f769-9a31-4555-be16-51e310ac34e3" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 13 08:47:15 crc kubenswrapper[4685]: I1013 08:47:15.556443 4685 patch_prober.go:28] interesting pod/router-default-5444994796-r96rq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 08:47:15 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Oct 13 08:47:15 crc kubenswrapper[4685]: [+]process-running ok Oct 13 08:47:15 crc kubenswrapper[4685]: healthz check failed Oct 13 08:47:15 crc kubenswrapper[4685]: I1013 08:47:15.556526 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r96rq" podUID="0fe547bf-ea91-4546-a886-c613abeae02a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 08:47:16 crc kubenswrapper[4685]: I1013 08:47:16.327559 4685 generic.go:334] "Generic (PLEG): container finished" podID="acf8bfed-51a6-4317-a60e-8b8ff193f47d" containerID="718c90afd70733759a927e4bd194fc54aca901f0a41f548c0772ee5cac31f74d" exitCode=0 Oct 13 08:47:16 crc kubenswrapper[4685]: I1013 08:47:16.327613 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"acf8bfed-51a6-4317-a60e-8b8ff193f47d","Type":"ContainerDied","Data":"718c90afd70733759a927e4bd194fc54aca901f0a41f548c0772ee5cac31f74d"} Oct 13 08:47:16 crc kubenswrapper[4685]: I1013 08:47:16.555291 4685 patch_prober.go:28] interesting pod/router-default-5444994796-r96rq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 08:47:16 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Oct 13 08:47:16 crc kubenswrapper[4685]: [+]process-running ok Oct 13 08:47:16 crc kubenswrapper[4685]: healthz check failed Oct 13 08:47:16 crc kubenswrapper[4685]: I1013 08:47:16.555356 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r96rq" podUID="0fe547bf-ea91-4546-a886-c613abeae02a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 08:47:16 crc kubenswrapper[4685]: I1013 08:47:16.873876 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" Oct 13 08:47:17 crc kubenswrapper[4685]: I1013 08:47:17.557393 4685 patch_prober.go:28] interesting pod/router-default-5444994796-r96rq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 08:47:17 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Oct 13 08:47:17 crc kubenswrapper[4685]: [+]process-running ok Oct 13 08:47:17 crc kubenswrapper[4685]: healthz check failed Oct 13 08:47:17 crc kubenswrapper[4685]: I1013 08:47:17.557513 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r96rq" podUID="0fe547bf-ea91-4546-a886-c613abeae02a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 08:47:17 crc kubenswrapper[4685]: I1013 08:47:17.862209 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 13 08:47:18 crc kubenswrapper[4685]: I1013 08:47:18.040688 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/acf8bfed-51a6-4317-a60e-8b8ff193f47d-kubelet-dir\") pod \"acf8bfed-51a6-4317-a60e-8b8ff193f47d\" (UID: \"acf8bfed-51a6-4317-a60e-8b8ff193f47d\") " Oct 13 08:47:18 crc kubenswrapper[4685]: I1013 08:47:18.040830 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/acf8bfed-51a6-4317-a60e-8b8ff193f47d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "acf8bfed-51a6-4317-a60e-8b8ff193f47d" (UID: "acf8bfed-51a6-4317-a60e-8b8ff193f47d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 08:47:18 crc kubenswrapper[4685]: I1013 08:47:18.040906 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/acf8bfed-51a6-4317-a60e-8b8ff193f47d-kube-api-access\") pod \"acf8bfed-51a6-4317-a60e-8b8ff193f47d\" (UID: \"acf8bfed-51a6-4317-a60e-8b8ff193f47d\") " Oct 13 08:47:18 crc kubenswrapper[4685]: I1013 08:47:18.041289 4685 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/acf8bfed-51a6-4317-a60e-8b8ff193f47d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 13 08:47:18 crc kubenswrapper[4685]: I1013 08:47:18.070243 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acf8bfed-51a6-4317-a60e-8b8ff193f47d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "acf8bfed-51a6-4317-a60e-8b8ff193f47d" (UID: "acf8bfed-51a6-4317-a60e-8b8ff193f47d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:47:18 crc kubenswrapper[4685]: I1013 08:47:18.142880 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/acf8bfed-51a6-4317-a60e-8b8ff193f47d-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 13 08:47:18 crc kubenswrapper[4685]: I1013 08:47:18.420495 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"acf8bfed-51a6-4317-a60e-8b8ff193f47d","Type":"ContainerDied","Data":"f2387814581fc423d7d6995f3b82f628e8e60ef7025ab894f2d3a1f0f558debc"} Oct 13 08:47:18 crc kubenswrapper[4685]: I1013 08:47:18.420979 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2387814581fc423d7d6995f3b82f628e8e60ef7025ab894f2d3a1f0f558debc" Oct 13 08:47:18 crc kubenswrapper[4685]: I1013 08:47:18.420551 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 13 08:47:18 crc kubenswrapper[4685]: I1013 08:47:18.555762 4685 patch_prober.go:28] interesting pod/router-default-5444994796-r96rq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 08:47:18 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Oct 13 08:47:18 crc kubenswrapper[4685]: [+]process-running ok Oct 13 08:47:18 crc kubenswrapper[4685]: healthz check failed Oct 13 08:47:18 crc kubenswrapper[4685]: I1013 08:47:18.556743 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r96rq" podUID="0fe547bf-ea91-4546-a886-c613abeae02a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 08:47:18 crc kubenswrapper[4685]: I1013 08:47:18.856191 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs\") pod \"network-metrics-daemon-nkxgw\" (UID: \"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\") " pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:47:18 crc kubenswrapper[4685]: I1013 08:47:18.862959 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cb40c32-d7d4-4a1b-9cda-1de0fab48566-metrics-certs\") pod \"network-metrics-daemon-nkxgw\" (UID: \"7cb40c32-d7d4-4a1b-9cda-1de0fab48566\") " pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:47:19 crc kubenswrapper[4685]: I1013 08:47:19.040241 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nkxgw" Oct 13 08:47:19 crc kubenswrapper[4685]: I1013 08:47:19.555544 4685 patch_prober.go:28] interesting pod/router-default-5444994796-r96rq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 08:47:19 crc kubenswrapper[4685]: [-]has-synced failed: reason withheld Oct 13 08:47:19 crc kubenswrapper[4685]: [+]process-running ok Oct 13 08:47:19 crc kubenswrapper[4685]: healthz check failed Oct 13 08:47:19 crc kubenswrapper[4685]: I1013 08:47:19.555660 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r96rq" podUID="0fe547bf-ea91-4546-a886-c613abeae02a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 08:47:20 crc kubenswrapper[4685]: I1013 08:47:20.555061 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-r96rq" Oct 13 08:47:20 crc kubenswrapper[4685]: I1013 08:47:20.559523 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-r96rq" Oct 13 08:47:22 crc kubenswrapper[4685]: I1013 08:47:22.980262 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 08:47:22 crc kubenswrapper[4685]: I1013 08:47:22.980734 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 08:47:25 crc kubenswrapper[4685]: I1013 08:47:25.004333 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:47:25 crc kubenswrapper[4685]: I1013 08:47:25.009554 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:47:25 crc kubenswrapper[4685]: I1013 08:47:25.257951 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-2kbsj" Oct 13 08:47:30 crc kubenswrapper[4685]: I1013 08:47:30.350888 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:47:36 crc kubenswrapper[4685]: I1013 08:47:36.860516 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b7nbg" Oct 13 08:47:43 crc kubenswrapper[4685]: I1013 08:47:43.733092 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 08:47:43 crc kubenswrapper[4685]: E1013 08:47:43.942799 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 13 08:47:43 crc kubenswrapper[4685]: E1013 08:47:43.943595 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rf74x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-wbzsb_openshift-marketplace(1ca80dd4-6294-4263-b938-188c17814556): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 13 08:47:43 crc kubenswrapper[4685]: E1013 08:47:43.944855 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-wbzsb" podUID="1ca80dd4-6294-4263-b938-188c17814556" Oct 13 08:47:43 crc kubenswrapper[4685]: E1013 08:47:43.953773 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 13 08:47:43 crc kubenswrapper[4685]: E1013 08:47:43.953958 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-btnt9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-wmkhb_openshift-marketplace(8beb316d-efe3-427b-a377-c84063b2b5d6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 13 08:47:43 crc kubenswrapper[4685]: E1013 08:47:43.955205 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-wmkhb" podUID="8beb316d-efe3-427b-a377-c84063b2b5d6" Oct 13 08:47:44 crc kubenswrapper[4685]: I1013 08:47:44.353449 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-nkxgw"] Oct 13 08:47:44 crc kubenswrapper[4685]: W1013 08:47:44.367873 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7cb40c32_d7d4_4a1b_9cda_1de0fab48566.slice/crio-3fe5762e58cf5baa9942a78f7a516cbb93e45eb9f4d4d2104b60432d1b31256e WatchSource:0}: Error finding container 3fe5762e58cf5baa9942a78f7a516cbb93e45eb9f4d4d2104b60432d1b31256e: Status 404 returned error can't find the container with id 3fe5762e58cf5baa9942a78f7a516cbb93e45eb9f4d4d2104b60432d1b31256e Oct 13 08:47:44 crc kubenswrapper[4685]: I1013 08:47:44.655743 4685 generic.go:334] "Generic (PLEG): container finished" podID="e74b537e-de07-4a47-85e7-e2bd5a72a11c" containerID="3c437e73337675f69ec91870d24f037a180a0a577b56a2a3e093d134d92b6643" exitCode=0 Oct 13 08:47:44 crc kubenswrapper[4685]: I1013 08:47:44.655849 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fmmfz" event={"ID":"e74b537e-de07-4a47-85e7-e2bd5a72a11c","Type":"ContainerDied","Data":"3c437e73337675f69ec91870d24f037a180a0a577b56a2a3e093d134d92b6643"} Oct 13 08:47:44 crc kubenswrapper[4685]: I1013 08:47:44.658146 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-nkxgw" event={"ID":"7cb40c32-d7d4-4a1b-9cda-1de0fab48566","Type":"ContainerStarted","Data":"3fe5762e58cf5baa9942a78f7a516cbb93e45eb9f4d4d2104b60432d1b31256e"} Oct 13 08:47:44 crc kubenswrapper[4685]: I1013 08:47:44.662663 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6zv9" event={"ID":"75521184-74d3-44b9-a410-b0b0fafa4384","Type":"ContainerStarted","Data":"dcefc1b2b3e483363f52e5e450854eec214d122c593dcd06649fb752bb8bca77"} Oct 13 08:47:44 crc kubenswrapper[4685]: I1013 08:47:44.682739 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2xf4" event={"ID":"c18864ff-4530-4655-b491-25f85b5604b4","Type":"ContainerStarted","Data":"f9bd1ebb227c6d6983a0124247db61e6eee4b6ccfaae43c714c12a503b5d1f81"} Oct 13 08:47:44 crc kubenswrapper[4685]: I1013 08:47:44.704349 4685 generic.go:334] "Generic (PLEG): container finished" podID="23545a99-b853-4a78-85a6-6186dc3bf903" containerID="4e07cceaa718b1733c04d9dca707713e6914cbc8e8d84f555b853f0ce507af46" exitCode=0 Oct 13 08:47:44 crc kubenswrapper[4685]: I1013 08:47:44.704609 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdbjj" event={"ID":"23545a99-b853-4a78-85a6-6186dc3bf903","Type":"ContainerDied","Data":"4e07cceaa718b1733c04d9dca707713e6914cbc8e8d84f555b853f0ce507af46"} Oct 13 08:47:44 crc kubenswrapper[4685]: I1013 08:47:44.728682 4685 generic.go:334] "Generic (PLEG): container finished" podID="2a006d49-7467-4c6a-9024-09de1da2d8da" containerID="c511010569ecc0d075a8abf8fa4f6711fde6a62953387de15b44be6828ea6ff0" exitCode=0 Oct 13 08:47:44 crc kubenswrapper[4685]: I1013 08:47:44.729045 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csdt2" event={"ID":"2a006d49-7467-4c6a-9024-09de1da2d8da","Type":"ContainerDied","Data":"c511010569ecc0d075a8abf8fa4f6711fde6a62953387de15b44be6828ea6ff0"} Oct 13 08:47:44 crc kubenswrapper[4685]: I1013 08:47:44.733199 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnb5g" event={"ID":"9a9fb57f-9b71-4922-aa37-2661e199871d","Type":"ContainerStarted","Data":"0bcd527e1bac8897e288800cec08d22fa55e1fc5f2ab1b6989dc464880781441"} Oct 13 08:47:44 crc kubenswrapper[4685]: E1013 08:47:44.750392 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-wbzsb" podUID="1ca80dd4-6294-4263-b938-188c17814556" Oct 13 08:47:44 crc kubenswrapper[4685]: E1013 08:47:44.755850 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wmkhb" podUID="8beb316d-efe3-427b-a377-c84063b2b5d6" Oct 13 08:47:45 crc kubenswrapper[4685]: I1013 08:47:45.755980 4685 generic.go:334] "Generic (PLEG): container finished" podID="9a9fb57f-9b71-4922-aa37-2661e199871d" containerID="0bcd527e1bac8897e288800cec08d22fa55e1fc5f2ab1b6989dc464880781441" exitCode=0 Oct 13 08:47:45 crc kubenswrapper[4685]: I1013 08:47:45.756174 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnb5g" event={"ID":"9a9fb57f-9b71-4922-aa37-2661e199871d","Type":"ContainerDied","Data":"0bcd527e1bac8897e288800cec08d22fa55e1fc5f2ab1b6989dc464880781441"} Oct 13 08:47:45 crc kubenswrapper[4685]: I1013 08:47:45.762015 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-nkxgw" event={"ID":"7cb40c32-d7d4-4a1b-9cda-1de0fab48566","Type":"ContainerStarted","Data":"f050ac7e1ffda5027778f618e4d79601133fe8f552a07f4dfce2c474052a7ea5"} Oct 13 08:47:45 crc kubenswrapper[4685]: I1013 08:47:45.762091 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-nkxgw" event={"ID":"7cb40c32-d7d4-4a1b-9cda-1de0fab48566","Type":"ContainerStarted","Data":"439c32491d442c8fa5865a593948d3453bf145ed728ebe75f5ed72c187296293"} Oct 13 08:47:45 crc kubenswrapper[4685]: I1013 08:47:45.770257 4685 generic.go:334] "Generic (PLEG): container finished" podID="75521184-74d3-44b9-a410-b0b0fafa4384" containerID="dcefc1b2b3e483363f52e5e450854eec214d122c593dcd06649fb752bb8bca77" exitCode=0 Oct 13 08:47:45 crc kubenswrapper[4685]: I1013 08:47:45.770353 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6zv9" event={"ID":"75521184-74d3-44b9-a410-b0b0fafa4384","Type":"ContainerDied","Data":"dcefc1b2b3e483363f52e5e450854eec214d122c593dcd06649fb752bb8bca77"} Oct 13 08:47:45 crc kubenswrapper[4685]: I1013 08:47:45.787755 4685 generic.go:334] "Generic (PLEG): container finished" podID="c18864ff-4530-4655-b491-25f85b5604b4" containerID="f9bd1ebb227c6d6983a0124247db61e6eee4b6ccfaae43c714c12a503b5d1f81" exitCode=0 Oct 13 08:47:45 crc kubenswrapper[4685]: I1013 08:47:45.787829 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2xf4" event={"ID":"c18864ff-4530-4655-b491-25f85b5604b4","Type":"ContainerDied","Data":"f9bd1ebb227c6d6983a0124247db61e6eee4b6ccfaae43c714c12a503b5d1f81"} Oct 13 08:47:45 crc kubenswrapper[4685]: I1013 08:47:45.841090 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-nkxgw" podStartSLOduration=169.841066291 podStartE2EDuration="2m49.841066291s" podCreationTimestamp="2025-10-13 08:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:47:45.839404922 +0000 UTC m=+190.987280703" watchObservedRunningTime="2025-10-13 08:47:45.841066291 +0000 UTC m=+190.988942062" Oct 13 08:47:47 crc kubenswrapper[4685]: I1013 08:47:47.806514 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2xf4" event={"ID":"c18864ff-4530-4655-b491-25f85b5604b4","Type":"ContainerStarted","Data":"1a119908677e9e402183a56e80d94719f3e996630383093eaa52f6f6443e1dc1"} Oct 13 08:47:47 crc kubenswrapper[4685]: I1013 08:47:47.833627 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-w2xf4" podStartSLOduration=4.843190855 podStartE2EDuration="42.833603482s" podCreationTimestamp="2025-10-13 08:47:05 +0000 UTC" firstStartedPulling="2025-10-13 08:47:08.62266709 +0000 UTC m=+153.770542851" lastFinishedPulling="2025-10-13 08:47:46.613079717 +0000 UTC m=+191.760955478" observedRunningTime="2025-10-13 08:47:47.83048514 +0000 UTC m=+192.978360941" watchObservedRunningTime="2025-10-13 08:47:47.833603482 +0000 UTC m=+192.981479243" Oct 13 08:47:48 crc kubenswrapper[4685]: I1013 08:47:48.819726 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fmmfz" event={"ID":"e74b537e-de07-4a47-85e7-e2bd5a72a11c","Type":"ContainerStarted","Data":"c743e565b6746e20610f8f71c8279e78db342f92df329e8b719fa201fafe481c"} Oct 13 08:47:49 crc kubenswrapper[4685]: I1013 08:47:49.826354 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csdt2" event={"ID":"2a006d49-7467-4c6a-9024-09de1da2d8da","Type":"ContainerStarted","Data":"e5747f16bfac75b031de8e8f2d09394881f44238c542df0f841d8ee83c74a6ab"} Oct 13 08:47:49 crc kubenswrapper[4685]: I1013 08:47:49.854288 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fmmfz" podStartSLOduration=4.714118937 podStartE2EDuration="42.854264936s" podCreationTimestamp="2025-10-13 08:47:07 +0000 UTC" firstStartedPulling="2025-10-13 08:47:09.953183352 +0000 UTC m=+155.101059103" lastFinishedPulling="2025-10-13 08:47:48.093329341 +0000 UTC m=+193.241205102" observedRunningTime="2025-10-13 08:47:48.846580182 +0000 UTC m=+193.994455993" watchObservedRunningTime="2025-10-13 08:47:49.854264936 +0000 UTC m=+195.002140697" Oct 13 08:47:49 crc kubenswrapper[4685]: I1013 08:47:49.857170 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-csdt2" podStartSLOduration=6.322537572 podStartE2EDuration="44.857156341s" podCreationTimestamp="2025-10-13 08:47:05 +0000 UTC" firstStartedPulling="2025-10-13 08:47:10.031203381 +0000 UTC m=+155.179079142" lastFinishedPulling="2025-10-13 08:47:48.56582215 +0000 UTC m=+193.713697911" observedRunningTime="2025-10-13 08:47:49.853233776 +0000 UTC m=+195.001109537" watchObservedRunningTime="2025-10-13 08:47:49.857156341 +0000 UTC m=+195.005032102" Oct 13 08:47:50 crc kubenswrapper[4685]: I1013 08:47:50.838474 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnb5g" event={"ID":"9a9fb57f-9b71-4922-aa37-2661e199871d","Type":"ContainerStarted","Data":"717469beb1aec88ed27860570bbc5b12178e61d213c7b483327ba207bf6dd169"} Oct 13 08:47:50 crc kubenswrapper[4685]: I1013 08:47:50.844154 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6zv9" event={"ID":"75521184-74d3-44b9-a410-b0b0fafa4384","Type":"ContainerStarted","Data":"14cd7f4d36a1d55aadc343b5a23b83f99053aa71adbfaf0b71f35ba7ccc783c9"} Oct 13 08:47:50 crc kubenswrapper[4685]: I1013 08:47:50.851318 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdbjj" event={"ID":"23545a99-b853-4a78-85a6-6186dc3bf903","Type":"ContainerStarted","Data":"6941be18324ed1a356f9c714f087aad7dd4be8224e0c1fc1ba27f4a5d5c846f7"} Oct 13 08:47:50 crc kubenswrapper[4685]: I1013 08:47:50.892937 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bnb5g" podStartSLOduration=2.454678189 podStartE2EDuration="42.892890596s" podCreationTimestamp="2025-10-13 08:47:08 +0000 UTC" firstStartedPulling="2025-10-13 08:47:09.921573185 +0000 UTC m=+155.069448946" lastFinishedPulling="2025-10-13 08:47:50.359785592 +0000 UTC m=+195.507661353" observedRunningTime="2025-10-13 08:47:50.870672428 +0000 UTC m=+196.018548189" watchObservedRunningTime="2025-10-13 08:47:50.892890596 +0000 UTC m=+196.040766357" Oct 13 08:47:50 crc kubenswrapper[4685]: I1013 08:47:50.895719 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vdbjj" podStartSLOduration=5.203117285 podStartE2EDuration="46.895707539s" podCreationTimestamp="2025-10-13 08:47:04 +0000 UTC" firstStartedPulling="2025-10-13 08:47:08.63178725 +0000 UTC m=+153.779663011" lastFinishedPulling="2025-10-13 08:47:50.324377504 +0000 UTC m=+195.472253265" observedRunningTime="2025-10-13 08:47:50.891680339 +0000 UTC m=+196.039556100" watchObservedRunningTime="2025-10-13 08:47:50.895707539 +0000 UTC m=+196.043583300" Oct 13 08:47:50 crc kubenswrapper[4685]: I1013 08:47:50.931047 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-v6zv9" podStartSLOduration=3.052389964 podStartE2EDuration="42.931024904s" podCreationTimestamp="2025-10-13 08:47:08 +0000 UTC" firstStartedPulling="2025-10-13 08:47:09.986779566 +0000 UTC m=+155.134655327" lastFinishedPulling="2025-10-13 08:47:49.865414506 +0000 UTC m=+195.013290267" observedRunningTime="2025-10-13 08:47:50.927076878 +0000 UTC m=+196.074952639" watchObservedRunningTime="2025-10-13 08:47:50.931024904 +0000 UTC m=+196.078900665" Oct 13 08:47:52 crc kubenswrapper[4685]: I1013 08:47:52.979839 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 08:47:52 crc kubenswrapper[4685]: I1013 08:47:52.979953 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 08:47:55 crc kubenswrapper[4685]: I1013 08:47:55.650385 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vdbjj" Oct 13 08:47:55 crc kubenswrapper[4685]: I1013 08:47:55.651262 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vdbjj" Oct 13 08:47:55 crc kubenswrapper[4685]: I1013 08:47:55.765444 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-w2xf4" Oct 13 08:47:55 crc kubenswrapper[4685]: I1013 08:47:55.766232 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-w2xf4" Oct 13 08:47:56 crc kubenswrapper[4685]: I1013 08:47:56.010719 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-csdt2" Oct 13 08:47:56 crc kubenswrapper[4685]: I1013 08:47:56.012423 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-csdt2" Oct 13 08:47:56 crc kubenswrapper[4685]: I1013 08:47:56.300810 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-w2xf4" Oct 13 08:47:56 crc kubenswrapper[4685]: I1013 08:47:56.317969 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-csdt2" Oct 13 08:47:56 crc kubenswrapper[4685]: I1013 08:47:56.318072 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vdbjj" Oct 13 08:47:56 crc kubenswrapper[4685]: I1013 08:47:56.385059 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vdbjj" Oct 13 08:47:56 crc kubenswrapper[4685]: I1013 08:47:56.931563 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-csdt2" Oct 13 08:47:56 crc kubenswrapper[4685]: I1013 08:47:56.942063 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-w2xf4" Oct 13 08:47:57 crc kubenswrapper[4685]: I1013 08:47:57.562145 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fmmfz" Oct 13 08:47:57 crc kubenswrapper[4685]: I1013 08:47:57.568772 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fmmfz" Oct 13 08:47:57 crc kubenswrapper[4685]: I1013 08:47:57.632142 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fmmfz" Oct 13 08:47:58 crc kubenswrapper[4685]: I1013 08:47:58.071228 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fmmfz" Oct 13 08:47:58 crc kubenswrapper[4685]: I1013 08:47:58.658288 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-v6zv9" Oct 13 08:47:58 crc kubenswrapper[4685]: I1013 08:47:58.658727 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v6zv9" Oct 13 08:47:58 crc kubenswrapper[4685]: I1013 08:47:58.708154 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-csdt2"] Oct 13 08:47:58 crc kubenswrapper[4685]: I1013 08:47:58.734341 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-v6zv9" Oct 13 08:47:58 crc kubenswrapper[4685]: I1013 08:47:58.905337 4685 generic.go:334] "Generic (PLEG): container finished" podID="8beb316d-efe3-427b-a377-c84063b2b5d6" containerID="66986a1f21bb4b3d2889a99e24dbcc5684512468a1a0cbb5767c4679c5f3beff" exitCode=0 Oct 13 08:47:58 crc kubenswrapper[4685]: I1013 08:47:58.905430 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmkhb" event={"ID":"8beb316d-efe3-427b-a377-c84063b2b5d6","Type":"ContainerDied","Data":"66986a1f21bb4b3d2889a99e24dbcc5684512468a1a0cbb5767c4679c5f3beff"} Oct 13 08:47:58 crc kubenswrapper[4685]: I1013 08:47:58.910828 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbzsb" event={"ID":"1ca80dd4-6294-4263-b938-188c17814556","Type":"ContainerStarted","Data":"6942d52617cb26bd17e1bd92a76ce68ff8e2c5cb68cdefb33f1e0da6d43f8963"} Oct 13 08:47:58 crc kubenswrapper[4685]: I1013 08:47:58.972077 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-v6zv9" Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.005269 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bnb5g" Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.005883 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bnb5g" Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.063366 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bnb5g" Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.306445 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w2xf4"] Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.307416 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-w2xf4" podUID="c18864ff-4530-4655-b491-25f85b5604b4" containerName="registry-server" containerID="cri-o://1a119908677e9e402183a56e80d94719f3e996630383093eaa52f6f6443e1dc1" gracePeriod=2 Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.733418 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w2xf4" Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.821372 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c18864ff-4530-4655-b491-25f85b5604b4-catalog-content\") pod \"c18864ff-4530-4655-b491-25f85b5604b4\" (UID: \"c18864ff-4530-4655-b491-25f85b5604b4\") " Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.821480 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4lpc\" (UniqueName: \"kubernetes.io/projected/c18864ff-4530-4655-b491-25f85b5604b4-kube-api-access-c4lpc\") pod \"c18864ff-4530-4655-b491-25f85b5604b4\" (UID: \"c18864ff-4530-4655-b491-25f85b5604b4\") " Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.821596 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c18864ff-4530-4655-b491-25f85b5604b4-utilities\") pod \"c18864ff-4530-4655-b491-25f85b5604b4\" (UID: \"c18864ff-4530-4655-b491-25f85b5604b4\") " Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.822720 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c18864ff-4530-4655-b491-25f85b5604b4-utilities" (OuterVolumeSpecName: "utilities") pod "c18864ff-4530-4655-b491-25f85b5604b4" (UID: "c18864ff-4530-4655-b491-25f85b5604b4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.830174 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c18864ff-4530-4655-b491-25f85b5604b4-kube-api-access-c4lpc" (OuterVolumeSpecName: "kube-api-access-c4lpc") pod "c18864ff-4530-4655-b491-25f85b5604b4" (UID: "c18864ff-4530-4655-b491-25f85b5604b4"). InnerVolumeSpecName "kube-api-access-c4lpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.881224 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c18864ff-4530-4655-b491-25f85b5604b4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c18864ff-4530-4655-b491-25f85b5604b4" (UID: "c18864ff-4530-4655-b491-25f85b5604b4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.923026 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c18864ff-4530-4655-b491-25f85b5604b4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.923063 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4lpc\" (UniqueName: \"kubernetes.io/projected/c18864ff-4530-4655-b491-25f85b5604b4-kube-api-access-c4lpc\") on node \"crc\" DevicePath \"\"" Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.923077 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c18864ff-4530-4655-b491-25f85b5604b4-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.923611 4685 generic.go:334] "Generic (PLEG): container finished" podID="c18864ff-4530-4655-b491-25f85b5604b4" containerID="1a119908677e9e402183a56e80d94719f3e996630383093eaa52f6f6443e1dc1" exitCode=0 Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.923701 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2xf4" event={"ID":"c18864ff-4530-4655-b491-25f85b5604b4","Type":"ContainerDied","Data":"1a119908677e9e402183a56e80d94719f3e996630383093eaa52f6f6443e1dc1"} Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.923744 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2xf4" event={"ID":"c18864ff-4530-4655-b491-25f85b5604b4","Type":"ContainerDied","Data":"d16dc906b086440cea6b3ceb1f6f831bd740160c3f79845aa559fbc6d7efa573"} Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.923767 4685 scope.go:117] "RemoveContainer" containerID="1a119908677e9e402183a56e80d94719f3e996630383093eaa52f6f6443e1dc1" Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.923968 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w2xf4" Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.929059 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmkhb" event={"ID":"8beb316d-efe3-427b-a377-c84063b2b5d6","Type":"ContainerStarted","Data":"c87b34dd8a191222824fd1e6415912af37812b2d16aa0a100f43151ddf9b38fe"} Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.936552 4685 generic.go:334] "Generic (PLEG): container finished" podID="1ca80dd4-6294-4263-b938-188c17814556" containerID="6942d52617cb26bd17e1bd92a76ce68ff8e2c5cb68cdefb33f1e0da6d43f8963" exitCode=0 Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.937144 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbzsb" event={"ID":"1ca80dd4-6294-4263-b938-188c17814556","Type":"ContainerDied","Data":"6942d52617cb26bd17e1bd92a76ce68ff8e2c5cb68cdefb33f1e0da6d43f8963"} Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.938620 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-csdt2" podUID="2a006d49-7467-4c6a-9024-09de1da2d8da" containerName="registry-server" containerID="cri-o://e5747f16bfac75b031de8e8f2d09394881f44238c542df0f841d8ee83c74a6ab" gracePeriod=2 Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.945439 4685 scope.go:117] "RemoveContainer" containerID="f9bd1ebb227c6d6983a0124247db61e6eee4b6ccfaae43c714c12a503b5d1f81" Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.979935 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wmkhb" podStartSLOduration=3.471163949 podStartE2EDuration="52.979890745s" podCreationTimestamp="2025-10-13 08:47:07 +0000 UTC" firstStartedPulling="2025-10-13 08:47:09.894774532 +0000 UTC m=+155.042650283" lastFinishedPulling="2025-10-13 08:47:59.403501298 +0000 UTC m=+204.551377079" observedRunningTime="2025-10-13 08:47:59.957410941 +0000 UTC m=+205.105286702" watchObservedRunningTime="2025-10-13 08:47:59.979890745 +0000 UTC m=+205.127766506" Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.983132 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w2xf4"] Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.989855 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-w2xf4"] Oct 13 08:47:59 crc kubenswrapper[4685]: I1013 08:47:59.996108 4685 scope.go:117] "RemoveContainer" containerID="b9f3050114c1ef462b6fcab1e3804bef5187bb91acf8d966c210310f3232a4fa" Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.028463 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bnb5g" Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.068254 4685 scope.go:117] "RemoveContainer" containerID="1a119908677e9e402183a56e80d94719f3e996630383093eaa52f6f6443e1dc1" Oct 13 08:48:00 crc kubenswrapper[4685]: E1013 08:48:00.068884 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a119908677e9e402183a56e80d94719f3e996630383093eaa52f6f6443e1dc1\": container with ID starting with 1a119908677e9e402183a56e80d94719f3e996630383093eaa52f6f6443e1dc1 not found: ID does not exist" containerID="1a119908677e9e402183a56e80d94719f3e996630383093eaa52f6f6443e1dc1" Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.069073 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a119908677e9e402183a56e80d94719f3e996630383093eaa52f6f6443e1dc1"} err="failed to get container status \"1a119908677e9e402183a56e80d94719f3e996630383093eaa52f6f6443e1dc1\": rpc error: code = NotFound desc = could not find container \"1a119908677e9e402183a56e80d94719f3e996630383093eaa52f6f6443e1dc1\": container with ID starting with 1a119908677e9e402183a56e80d94719f3e996630383093eaa52f6f6443e1dc1 not found: ID does not exist" Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.069215 4685 scope.go:117] "RemoveContainer" containerID="f9bd1ebb227c6d6983a0124247db61e6eee4b6ccfaae43c714c12a503b5d1f81" Oct 13 08:48:00 crc kubenswrapper[4685]: E1013 08:48:00.069748 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9bd1ebb227c6d6983a0124247db61e6eee4b6ccfaae43c714c12a503b5d1f81\": container with ID starting with f9bd1ebb227c6d6983a0124247db61e6eee4b6ccfaae43c714c12a503b5d1f81 not found: ID does not exist" containerID="f9bd1ebb227c6d6983a0124247db61e6eee4b6ccfaae43c714c12a503b5d1f81" Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.069809 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9bd1ebb227c6d6983a0124247db61e6eee4b6ccfaae43c714c12a503b5d1f81"} err="failed to get container status \"f9bd1ebb227c6d6983a0124247db61e6eee4b6ccfaae43c714c12a503b5d1f81\": rpc error: code = NotFound desc = could not find container \"f9bd1ebb227c6d6983a0124247db61e6eee4b6ccfaae43c714c12a503b5d1f81\": container with ID starting with f9bd1ebb227c6d6983a0124247db61e6eee4b6ccfaae43c714c12a503b5d1f81 not found: ID does not exist" Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.069848 4685 scope.go:117] "RemoveContainer" containerID="b9f3050114c1ef462b6fcab1e3804bef5187bb91acf8d966c210310f3232a4fa" Oct 13 08:48:00 crc kubenswrapper[4685]: E1013 08:48:00.070194 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9f3050114c1ef462b6fcab1e3804bef5187bb91acf8d966c210310f3232a4fa\": container with ID starting with b9f3050114c1ef462b6fcab1e3804bef5187bb91acf8d966c210310f3232a4fa not found: ID does not exist" containerID="b9f3050114c1ef462b6fcab1e3804bef5187bb91acf8d966c210310f3232a4fa" Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.070293 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9f3050114c1ef462b6fcab1e3804bef5187bb91acf8d966c210310f3232a4fa"} err="failed to get container status \"b9f3050114c1ef462b6fcab1e3804bef5187bb91acf8d966c210310f3232a4fa\": rpc error: code = NotFound desc = could not find container \"b9f3050114c1ef462b6fcab1e3804bef5187bb91acf8d966c210310f3232a4fa\": container with ID starting with b9f3050114c1ef462b6fcab1e3804bef5187bb91acf8d966c210310f3232a4fa not found: ID does not exist" Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.332577 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-csdt2" Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.431145 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a006d49-7467-4c6a-9024-09de1da2d8da-utilities\") pod \"2a006d49-7467-4c6a-9024-09de1da2d8da\" (UID: \"2a006d49-7467-4c6a-9024-09de1da2d8da\") " Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.431260 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a006d49-7467-4c6a-9024-09de1da2d8da-catalog-content\") pod \"2a006d49-7467-4c6a-9024-09de1da2d8da\" (UID: \"2a006d49-7467-4c6a-9024-09de1da2d8da\") " Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.431309 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jch9n\" (UniqueName: \"kubernetes.io/projected/2a006d49-7467-4c6a-9024-09de1da2d8da-kube-api-access-jch9n\") pod \"2a006d49-7467-4c6a-9024-09de1da2d8da\" (UID: \"2a006d49-7467-4c6a-9024-09de1da2d8da\") " Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.432143 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a006d49-7467-4c6a-9024-09de1da2d8da-utilities" (OuterVolumeSpecName: "utilities") pod "2a006d49-7467-4c6a-9024-09de1da2d8da" (UID: "2a006d49-7467-4c6a-9024-09de1da2d8da"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.448443 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a006d49-7467-4c6a-9024-09de1da2d8da-kube-api-access-jch9n" (OuterVolumeSpecName: "kube-api-access-jch9n") pod "2a006d49-7467-4c6a-9024-09de1da2d8da" (UID: "2a006d49-7467-4c6a-9024-09de1da2d8da"). InnerVolumeSpecName "kube-api-access-jch9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.486871 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a006d49-7467-4c6a-9024-09de1da2d8da-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2a006d49-7467-4c6a-9024-09de1da2d8da" (UID: "2a006d49-7467-4c6a-9024-09de1da2d8da"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.533669 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a006d49-7467-4c6a-9024-09de1da2d8da-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.533726 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a006d49-7467-4c6a-9024-09de1da2d8da-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.533746 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jch9n\" (UniqueName: \"kubernetes.io/projected/2a006d49-7467-4c6a-9024-09de1da2d8da-kube-api-access-jch9n\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.946213 4685 generic.go:334] "Generic (PLEG): container finished" podID="2a006d49-7467-4c6a-9024-09de1da2d8da" containerID="e5747f16bfac75b031de8e8f2d09394881f44238c542df0f841d8ee83c74a6ab" exitCode=0 Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.946421 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csdt2" event={"ID":"2a006d49-7467-4c6a-9024-09de1da2d8da","Type":"ContainerDied","Data":"e5747f16bfac75b031de8e8f2d09394881f44238c542df0f841d8ee83c74a6ab"} Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.947315 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csdt2" event={"ID":"2a006d49-7467-4c6a-9024-09de1da2d8da","Type":"ContainerDied","Data":"46a94fdcf5d71b5172f76dd666d6bca2896e9c295a66888988381093d17fee2e"} Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.947438 4685 scope.go:117] "RemoveContainer" containerID="e5747f16bfac75b031de8e8f2d09394881f44238c542df0f841d8ee83c74a6ab" Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.946522 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-csdt2" Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.968944 4685 scope.go:117] "RemoveContainer" containerID="c511010569ecc0d075a8abf8fa4f6711fde6a62953387de15b44be6828ea6ff0" Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.986793 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-csdt2"] Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.991621 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-csdt2"] Oct 13 08:48:00 crc kubenswrapper[4685]: I1013 08:48:00.998614 4685 scope.go:117] "RemoveContainer" containerID="996e1c2258995f0e196833ae4d02fed1511f882f91438ac534fe16b223f11b14" Oct 13 08:48:01 crc kubenswrapper[4685]: I1013 08:48:01.018596 4685 scope.go:117] "RemoveContainer" containerID="e5747f16bfac75b031de8e8f2d09394881f44238c542df0f841d8ee83c74a6ab" Oct 13 08:48:01 crc kubenswrapper[4685]: E1013 08:48:01.019447 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5747f16bfac75b031de8e8f2d09394881f44238c542df0f841d8ee83c74a6ab\": container with ID starting with e5747f16bfac75b031de8e8f2d09394881f44238c542df0f841d8ee83c74a6ab not found: ID does not exist" containerID="e5747f16bfac75b031de8e8f2d09394881f44238c542df0f841d8ee83c74a6ab" Oct 13 08:48:01 crc kubenswrapper[4685]: I1013 08:48:01.019579 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5747f16bfac75b031de8e8f2d09394881f44238c542df0f841d8ee83c74a6ab"} err="failed to get container status \"e5747f16bfac75b031de8e8f2d09394881f44238c542df0f841d8ee83c74a6ab\": rpc error: code = NotFound desc = could not find container \"e5747f16bfac75b031de8e8f2d09394881f44238c542df0f841d8ee83c74a6ab\": container with ID starting with e5747f16bfac75b031de8e8f2d09394881f44238c542df0f841d8ee83c74a6ab not found: ID does not exist" Oct 13 08:48:01 crc kubenswrapper[4685]: I1013 08:48:01.019670 4685 scope.go:117] "RemoveContainer" containerID="c511010569ecc0d075a8abf8fa4f6711fde6a62953387de15b44be6828ea6ff0" Oct 13 08:48:01 crc kubenswrapper[4685]: E1013 08:48:01.020174 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c511010569ecc0d075a8abf8fa4f6711fde6a62953387de15b44be6828ea6ff0\": container with ID starting with c511010569ecc0d075a8abf8fa4f6711fde6a62953387de15b44be6828ea6ff0 not found: ID does not exist" containerID="c511010569ecc0d075a8abf8fa4f6711fde6a62953387de15b44be6828ea6ff0" Oct 13 08:48:01 crc kubenswrapper[4685]: I1013 08:48:01.020267 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c511010569ecc0d075a8abf8fa4f6711fde6a62953387de15b44be6828ea6ff0"} err="failed to get container status \"c511010569ecc0d075a8abf8fa4f6711fde6a62953387de15b44be6828ea6ff0\": rpc error: code = NotFound desc = could not find container \"c511010569ecc0d075a8abf8fa4f6711fde6a62953387de15b44be6828ea6ff0\": container with ID starting with c511010569ecc0d075a8abf8fa4f6711fde6a62953387de15b44be6828ea6ff0 not found: ID does not exist" Oct 13 08:48:01 crc kubenswrapper[4685]: I1013 08:48:01.020341 4685 scope.go:117] "RemoveContainer" containerID="996e1c2258995f0e196833ae4d02fed1511f882f91438ac534fe16b223f11b14" Oct 13 08:48:01 crc kubenswrapper[4685]: E1013 08:48:01.020687 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"996e1c2258995f0e196833ae4d02fed1511f882f91438ac534fe16b223f11b14\": container with ID starting with 996e1c2258995f0e196833ae4d02fed1511f882f91438ac534fe16b223f11b14 not found: ID does not exist" containerID="996e1c2258995f0e196833ae4d02fed1511f882f91438ac534fe16b223f11b14" Oct 13 08:48:01 crc kubenswrapper[4685]: I1013 08:48:01.020788 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"996e1c2258995f0e196833ae4d02fed1511f882f91438ac534fe16b223f11b14"} err="failed to get container status \"996e1c2258995f0e196833ae4d02fed1511f882f91438ac534fe16b223f11b14\": rpc error: code = NotFound desc = could not find container \"996e1c2258995f0e196833ae4d02fed1511f882f91438ac534fe16b223f11b14\": container with ID starting with 996e1c2258995f0e196833ae4d02fed1511f882f91438ac534fe16b223f11b14 not found: ID does not exist" Oct 13 08:48:01 crc kubenswrapper[4685]: I1013 08:48:01.509738 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a006d49-7467-4c6a-9024-09de1da2d8da" path="/var/lib/kubelet/pods/2a006d49-7467-4c6a-9024-09de1da2d8da/volumes" Oct 13 08:48:01 crc kubenswrapper[4685]: I1013 08:48:01.510380 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c18864ff-4530-4655-b491-25f85b5604b4" path="/var/lib/kubelet/pods/c18864ff-4530-4655-b491-25f85b5604b4/volumes" Oct 13 08:48:01 crc kubenswrapper[4685]: I1013 08:48:01.963961 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbzsb" event={"ID":"1ca80dd4-6294-4263-b938-188c17814556","Type":"ContainerStarted","Data":"f001baad0e6618d53b2401c1e76f5498a24e3b3c4fa88b92f6b681279337490f"} Oct 13 08:48:01 crc kubenswrapper[4685]: I1013 08:48:01.993372 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wbzsb" podStartSLOduration=4.852463213 podStartE2EDuration="56.993351539s" podCreationTimestamp="2025-10-13 08:47:05 +0000 UTC" firstStartedPulling="2025-10-13 08:47:08.691685553 +0000 UTC m=+153.839561324" lastFinishedPulling="2025-10-13 08:48:00.832573889 +0000 UTC m=+205.980449650" observedRunningTime="2025-10-13 08:48:01.99063593 +0000 UTC m=+207.138511691" watchObservedRunningTime="2025-10-13 08:48:01.993351539 +0000 UTC m=+207.141227310" Oct 13 08:48:03 crc kubenswrapper[4685]: I1013 08:48:03.106905 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bnb5g"] Oct 13 08:48:03 crc kubenswrapper[4685]: I1013 08:48:03.107551 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bnb5g" podUID="9a9fb57f-9b71-4922-aa37-2661e199871d" containerName="registry-server" containerID="cri-o://717469beb1aec88ed27860570bbc5b12178e61d213c7b483327ba207bf6dd169" gracePeriod=2 Oct 13 08:48:03 crc kubenswrapper[4685]: I1013 08:48:03.490007 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bnb5g" Oct 13 08:48:03 crc kubenswrapper[4685]: I1013 08:48:03.576078 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a9fb57f-9b71-4922-aa37-2661e199871d-utilities\") pod \"9a9fb57f-9b71-4922-aa37-2661e199871d\" (UID: \"9a9fb57f-9b71-4922-aa37-2661e199871d\") " Oct 13 08:48:03 crc kubenswrapper[4685]: I1013 08:48:03.576795 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a9fb57f-9b71-4922-aa37-2661e199871d-catalog-content\") pod \"9a9fb57f-9b71-4922-aa37-2661e199871d\" (UID: \"9a9fb57f-9b71-4922-aa37-2661e199871d\") " Oct 13 08:48:03 crc kubenswrapper[4685]: I1013 08:48:03.576901 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tjlv\" (UniqueName: \"kubernetes.io/projected/9a9fb57f-9b71-4922-aa37-2661e199871d-kube-api-access-2tjlv\") pod \"9a9fb57f-9b71-4922-aa37-2661e199871d\" (UID: \"9a9fb57f-9b71-4922-aa37-2661e199871d\") " Oct 13 08:48:03 crc kubenswrapper[4685]: I1013 08:48:03.577442 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a9fb57f-9b71-4922-aa37-2661e199871d-utilities" (OuterVolumeSpecName: "utilities") pod "9a9fb57f-9b71-4922-aa37-2661e199871d" (UID: "9a9fb57f-9b71-4922-aa37-2661e199871d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:48:03 crc kubenswrapper[4685]: I1013 08:48:03.578187 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a9fb57f-9b71-4922-aa37-2661e199871d-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:03 crc kubenswrapper[4685]: I1013 08:48:03.583171 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a9fb57f-9b71-4922-aa37-2661e199871d-kube-api-access-2tjlv" (OuterVolumeSpecName: "kube-api-access-2tjlv") pod "9a9fb57f-9b71-4922-aa37-2661e199871d" (UID: "9a9fb57f-9b71-4922-aa37-2661e199871d"). InnerVolumeSpecName "kube-api-access-2tjlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:48:03 crc kubenswrapper[4685]: I1013 08:48:03.668563 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a9fb57f-9b71-4922-aa37-2661e199871d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9a9fb57f-9b71-4922-aa37-2661e199871d" (UID: "9a9fb57f-9b71-4922-aa37-2661e199871d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:48:03 crc kubenswrapper[4685]: I1013 08:48:03.680798 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a9fb57f-9b71-4922-aa37-2661e199871d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:03 crc kubenswrapper[4685]: I1013 08:48:03.680851 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tjlv\" (UniqueName: \"kubernetes.io/projected/9a9fb57f-9b71-4922-aa37-2661e199871d-kube-api-access-2tjlv\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:03 crc kubenswrapper[4685]: I1013 08:48:03.978617 4685 generic.go:334] "Generic (PLEG): container finished" podID="9a9fb57f-9b71-4922-aa37-2661e199871d" containerID="717469beb1aec88ed27860570bbc5b12178e61d213c7b483327ba207bf6dd169" exitCode=0 Oct 13 08:48:03 crc kubenswrapper[4685]: I1013 08:48:03.978859 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnb5g" event={"ID":"9a9fb57f-9b71-4922-aa37-2661e199871d","Type":"ContainerDied","Data":"717469beb1aec88ed27860570bbc5b12178e61d213c7b483327ba207bf6dd169"} Oct 13 08:48:03 crc kubenswrapper[4685]: I1013 08:48:03.979259 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bnb5g" Oct 13 08:48:03 crc kubenswrapper[4685]: I1013 08:48:03.979387 4685 scope.go:117] "RemoveContainer" containerID="717469beb1aec88ed27860570bbc5b12178e61d213c7b483327ba207bf6dd169" Oct 13 08:48:03 crc kubenswrapper[4685]: I1013 08:48:03.979658 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnb5g" event={"ID":"9a9fb57f-9b71-4922-aa37-2661e199871d","Type":"ContainerDied","Data":"9cdf134f5276561e6ec5d70e073b1b8bf61059d01d473173797f4897637c0ea6"} Oct 13 08:48:04 crc kubenswrapper[4685]: I1013 08:48:04.006724 4685 scope.go:117] "RemoveContainer" containerID="0bcd527e1bac8897e288800cec08d22fa55e1fc5f2ab1b6989dc464880781441" Oct 13 08:48:04 crc kubenswrapper[4685]: I1013 08:48:04.023508 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bnb5g"] Oct 13 08:48:04 crc kubenswrapper[4685]: I1013 08:48:04.030900 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bnb5g"] Oct 13 08:48:04 crc kubenswrapper[4685]: I1013 08:48:04.063791 4685 scope.go:117] "RemoveContainer" containerID="c056ab57ddd7b0cf23825e2a8a185e5c971a0ec89eac9b83db07e0032008a567" Oct 13 08:48:04 crc kubenswrapper[4685]: I1013 08:48:04.091368 4685 scope.go:117] "RemoveContainer" containerID="717469beb1aec88ed27860570bbc5b12178e61d213c7b483327ba207bf6dd169" Oct 13 08:48:04 crc kubenswrapper[4685]: E1013 08:48:04.092300 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"717469beb1aec88ed27860570bbc5b12178e61d213c7b483327ba207bf6dd169\": container with ID starting with 717469beb1aec88ed27860570bbc5b12178e61d213c7b483327ba207bf6dd169 not found: ID does not exist" containerID="717469beb1aec88ed27860570bbc5b12178e61d213c7b483327ba207bf6dd169" Oct 13 08:48:04 crc kubenswrapper[4685]: I1013 08:48:04.092338 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"717469beb1aec88ed27860570bbc5b12178e61d213c7b483327ba207bf6dd169"} err="failed to get container status \"717469beb1aec88ed27860570bbc5b12178e61d213c7b483327ba207bf6dd169\": rpc error: code = NotFound desc = could not find container \"717469beb1aec88ed27860570bbc5b12178e61d213c7b483327ba207bf6dd169\": container with ID starting with 717469beb1aec88ed27860570bbc5b12178e61d213c7b483327ba207bf6dd169 not found: ID does not exist" Oct 13 08:48:04 crc kubenswrapper[4685]: I1013 08:48:04.092365 4685 scope.go:117] "RemoveContainer" containerID="0bcd527e1bac8897e288800cec08d22fa55e1fc5f2ab1b6989dc464880781441" Oct 13 08:48:04 crc kubenswrapper[4685]: E1013 08:48:04.092782 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bcd527e1bac8897e288800cec08d22fa55e1fc5f2ab1b6989dc464880781441\": container with ID starting with 0bcd527e1bac8897e288800cec08d22fa55e1fc5f2ab1b6989dc464880781441 not found: ID does not exist" containerID="0bcd527e1bac8897e288800cec08d22fa55e1fc5f2ab1b6989dc464880781441" Oct 13 08:48:04 crc kubenswrapper[4685]: I1013 08:48:04.092804 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bcd527e1bac8897e288800cec08d22fa55e1fc5f2ab1b6989dc464880781441"} err="failed to get container status \"0bcd527e1bac8897e288800cec08d22fa55e1fc5f2ab1b6989dc464880781441\": rpc error: code = NotFound desc = could not find container \"0bcd527e1bac8897e288800cec08d22fa55e1fc5f2ab1b6989dc464880781441\": container with ID starting with 0bcd527e1bac8897e288800cec08d22fa55e1fc5f2ab1b6989dc464880781441 not found: ID does not exist" Oct 13 08:48:04 crc kubenswrapper[4685]: I1013 08:48:04.092819 4685 scope.go:117] "RemoveContainer" containerID="c056ab57ddd7b0cf23825e2a8a185e5c971a0ec89eac9b83db07e0032008a567" Oct 13 08:48:04 crc kubenswrapper[4685]: E1013 08:48:04.093380 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c056ab57ddd7b0cf23825e2a8a185e5c971a0ec89eac9b83db07e0032008a567\": container with ID starting with c056ab57ddd7b0cf23825e2a8a185e5c971a0ec89eac9b83db07e0032008a567 not found: ID does not exist" containerID="c056ab57ddd7b0cf23825e2a8a185e5c971a0ec89eac9b83db07e0032008a567" Oct 13 08:48:04 crc kubenswrapper[4685]: I1013 08:48:04.093414 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c056ab57ddd7b0cf23825e2a8a185e5c971a0ec89eac9b83db07e0032008a567"} err="failed to get container status \"c056ab57ddd7b0cf23825e2a8a185e5c971a0ec89eac9b83db07e0032008a567\": rpc error: code = NotFound desc = could not find container \"c056ab57ddd7b0cf23825e2a8a185e5c971a0ec89eac9b83db07e0032008a567\": container with ID starting with c056ab57ddd7b0cf23825e2a8a185e5c971a0ec89eac9b83db07e0032008a567 not found: ID does not exist" Oct 13 08:48:05 crc kubenswrapper[4685]: I1013 08:48:05.250454 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-54dh8"] Oct 13 08:48:05 crc kubenswrapper[4685]: I1013 08:48:05.514491 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a9fb57f-9b71-4922-aa37-2661e199871d" path="/var/lib/kubelet/pods/9a9fb57f-9b71-4922-aa37-2661e199871d/volumes" Oct 13 08:48:05 crc kubenswrapper[4685]: I1013 08:48:05.527646 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wbzsb" Oct 13 08:48:05 crc kubenswrapper[4685]: I1013 08:48:05.527723 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wbzsb" Oct 13 08:48:05 crc kubenswrapper[4685]: I1013 08:48:05.570686 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wbzsb" Oct 13 08:48:07 crc kubenswrapper[4685]: I1013 08:48:07.986001 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wmkhb" Oct 13 08:48:07 crc kubenswrapper[4685]: I1013 08:48:07.986563 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wmkhb" Oct 13 08:48:08 crc kubenswrapper[4685]: I1013 08:48:08.024894 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wmkhb" Oct 13 08:48:08 crc kubenswrapper[4685]: I1013 08:48:08.066015 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wmkhb" Oct 13 08:48:09 crc kubenswrapper[4685]: I1013 08:48:09.910580 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wmkhb"] Oct 13 08:48:10 crc kubenswrapper[4685]: I1013 08:48:10.016807 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wmkhb" podUID="8beb316d-efe3-427b-a377-c84063b2b5d6" containerName="registry-server" containerID="cri-o://c87b34dd8a191222824fd1e6415912af37812b2d16aa0a100f43151ddf9b38fe" gracePeriod=2 Oct 13 08:48:10 crc kubenswrapper[4685]: I1013 08:48:10.387648 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wmkhb" Oct 13 08:48:10 crc kubenswrapper[4685]: I1013 08:48:10.487714 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8beb316d-efe3-427b-a377-c84063b2b5d6-catalog-content\") pod \"8beb316d-efe3-427b-a377-c84063b2b5d6\" (UID: \"8beb316d-efe3-427b-a377-c84063b2b5d6\") " Oct 13 08:48:10 crc kubenswrapper[4685]: I1013 08:48:10.487773 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8beb316d-efe3-427b-a377-c84063b2b5d6-utilities\") pod \"8beb316d-efe3-427b-a377-c84063b2b5d6\" (UID: \"8beb316d-efe3-427b-a377-c84063b2b5d6\") " Oct 13 08:48:10 crc kubenswrapper[4685]: I1013 08:48:10.487854 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btnt9\" (UniqueName: \"kubernetes.io/projected/8beb316d-efe3-427b-a377-c84063b2b5d6-kube-api-access-btnt9\") pod \"8beb316d-efe3-427b-a377-c84063b2b5d6\" (UID: \"8beb316d-efe3-427b-a377-c84063b2b5d6\") " Oct 13 08:48:10 crc kubenswrapper[4685]: I1013 08:48:10.489326 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8beb316d-efe3-427b-a377-c84063b2b5d6-utilities" (OuterVolumeSpecName: "utilities") pod "8beb316d-efe3-427b-a377-c84063b2b5d6" (UID: "8beb316d-efe3-427b-a377-c84063b2b5d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:48:10 crc kubenswrapper[4685]: I1013 08:48:10.495588 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8beb316d-efe3-427b-a377-c84063b2b5d6-kube-api-access-btnt9" (OuterVolumeSpecName: "kube-api-access-btnt9") pod "8beb316d-efe3-427b-a377-c84063b2b5d6" (UID: "8beb316d-efe3-427b-a377-c84063b2b5d6"). InnerVolumeSpecName "kube-api-access-btnt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:48:10 crc kubenswrapper[4685]: I1013 08:48:10.513268 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8beb316d-efe3-427b-a377-c84063b2b5d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8beb316d-efe3-427b-a377-c84063b2b5d6" (UID: "8beb316d-efe3-427b-a377-c84063b2b5d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:48:10 crc kubenswrapper[4685]: I1013 08:48:10.591176 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8beb316d-efe3-427b-a377-c84063b2b5d6-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:10 crc kubenswrapper[4685]: I1013 08:48:10.591233 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btnt9\" (UniqueName: \"kubernetes.io/projected/8beb316d-efe3-427b-a377-c84063b2b5d6-kube-api-access-btnt9\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:10 crc kubenswrapper[4685]: I1013 08:48:10.591246 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8beb316d-efe3-427b-a377-c84063b2b5d6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:11 crc kubenswrapper[4685]: I1013 08:48:11.026529 4685 generic.go:334] "Generic (PLEG): container finished" podID="8beb316d-efe3-427b-a377-c84063b2b5d6" containerID="c87b34dd8a191222824fd1e6415912af37812b2d16aa0a100f43151ddf9b38fe" exitCode=0 Oct 13 08:48:11 crc kubenswrapper[4685]: I1013 08:48:11.026619 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmkhb" event={"ID":"8beb316d-efe3-427b-a377-c84063b2b5d6","Type":"ContainerDied","Data":"c87b34dd8a191222824fd1e6415912af37812b2d16aa0a100f43151ddf9b38fe"} Oct 13 08:48:11 crc kubenswrapper[4685]: I1013 08:48:11.026688 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmkhb" event={"ID":"8beb316d-efe3-427b-a377-c84063b2b5d6","Type":"ContainerDied","Data":"034a890660d9f39279f8b27013e32a015e13bce374b32fec9d733d5db39758d2"} Oct 13 08:48:11 crc kubenswrapper[4685]: I1013 08:48:11.026708 4685 scope.go:117] "RemoveContainer" containerID="c87b34dd8a191222824fd1e6415912af37812b2d16aa0a100f43151ddf9b38fe" Oct 13 08:48:11 crc kubenswrapper[4685]: I1013 08:48:11.027957 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wmkhb" Oct 13 08:48:11 crc kubenswrapper[4685]: I1013 08:48:11.046166 4685 scope.go:117] "RemoveContainer" containerID="66986a1f21bb4b3d2889a99e24dbcc5684512468a1a0cbb5767c4679c5f3beff" Oct 13 08:48:11 crc kubenswrapper[4685]: I1013 08:48:11.062626 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wmkhb"] Oct 13 08:48:11 crc kubenswrapper[4685]: I1013 08:48:11.069977 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wmkhb"] Oct 13 08:48:11 crc kubenswrapper[4685]: I1013 08:48:11.085235 4685 scope.go:117] "RemoveContainer" containerID="085a1258cf27bb562aa13a64bd7c5df03a897160a7f63f1e37e21487aef8d60e" Oct 13 08:48:11 crc kubenswrapper[4685]: I1013 08:48:11.113050 4685 scope.go:117] "RemoveContainer" containerID="c87b34dd8a191222824fd1e6415912af37812b2d16aa0a100f43151ddf9b38fe" Oct 13 08:48:11 crc kubenswrapper[4685]: E1013 08:48:11.113702 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c87b34dd8a191222824fd1e6415912af37812b2d16aa0a100f43151ddf9b38fe\": container with ID starting with c87b34dd8a191222824fd1e6415912af37812b2d16aa0a100f43151ddf9b38fe not found: ID does not exist" containerID="c87b34dd8a191222824fd1e6415912af37812b2d16aa0a100f43151ddf9b38fe" Oct 13 08:48:11 crc kubenswrapper[4685]: I1013 08:48:11.113738 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c87b34dd8a191222824fd1e6415912af37812b2d16aa0a100f43151ddf9b38fe"} err="failed to get container status \"c87b34dd8a191222824fd1e6415912af37812b2d16aa0a100f43151ddf9b38fe\": rpc error: code = NotFound desc = could not find container \"c87b34dd8a191222824fd1e6415912af37812b2d16aa0a100f43151ddf9b38fe\": container with ID starting with c87b34dd8a191222824fd1e6415912af37812b2d16aa0a100f43151ddf9b38fe not found: ID does not exist" Oct 13 08:48:11 crc kubenswrapper[4685]: I1013 08:48:11.113779 4685 scope.go:117] "RemoveContainer" containerID="66986a1f21bb4b3d2889a99e24dbcc5684512468a1a0cbb5767c4679c5f3beff" Oct 13 08:48:11 crc kubenswrapper[4685]: E1013 08:48:11.114152 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66986a1f21bb4b3d2889a99e24dbcc5684512468a1a0cbb5767c4679c5f3beff\": container with ID starting with 66986a1f21bb4b3d2889a99e24dbcc5684512468a1a0cbb5767c4679c5f3beff not found: ID does not exist" containerID="66986a1f21bb4b3d2889a99e24dbcc5684512468a1a0cbb5767c4679c5f3beff" Oct 13 08:48:11 crc kubenswrapper[4685]: I1013 08:48:11.114173 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66986a1f21bb4b3d2889a99e24dbcc5684512468a1a0cbb5767c4679c5f3beff"} err="failed to get container status \"66986a1f21bb4b3d2889a99e24dbcc5684512468a1a0cbb5767c4679c5f3beff\": rpc error: code = NotFound desc = could not find container \"66986a1f21bb4b3d2889a99e24dbcc5684512468a1a0cbb5767c4679c5f3beff\": container with ID starting with 66986a1f21bb4b3d2889a99e24dbcc5684512468a1a0cbb5767c4679c5f3beff not found: ID does not exist" Oct 13 08:48:11 crc kubenswrapper[4685]: I1013 08:48:11.114187 4685 scope.go:117] "RemoveContainer" containerID="085a1258cf27bb562aa13a64bd7c5df03a897160a7f63f1e37e21487aef8d60e" Oct 13 08:48:11 crc kubenswrapper[4685]: E1013 08:48:11.114630 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"085a1258cf27bb562aa13a64bd7c5df03a897160a7f63f1e37e21487aef8d60e\": container with ID starting with 085a1258cf27bb562aa13a64bd7c5df03a897160a7f63f1e37e21487aef8d60e not found: ID does not exist" containerID="085a1258cf27bb562aa13a64bd7c5df03a897160a7f63f1e37e21487aef8d60e" Oct 13 08:48:11 crc kubenswrapper[4685]: I1013 08:48:11.114648 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"085a1258cf27bb562aa13a64bd7c5df03a897160a7f63f1e37e21487aef8d60e"} err="failed to get container status \"085a1258cf27bb562aa13a64bd7c5df03a897160a7f63f1e37e21487aef8d60e\": rpc error: code = NotFound desc = could not find container \"085a1258cf27bb562aa13a64bd7c5df03a897160a7f63f1e37e21487aef8d60e\": container with ID starting with 085a1258cf27bb562aa13a64bd7c5df03a897160a7f63f1e37e21487aef8d60e not found: ID does not exist" Oct 13 08:48:11 crc kubenswrapper[4685]: I1013 08:48:11.510963 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8beb316d-efe3-427b-a377-c84063b2b5d6" path="/var/lib/kubelet/pods/8beb316d-efe3-427b-a377-c84063b2b5d6/volumes" Oct 13 08:48:15 crc kubenswrapper[4685]: I1013 08:48:15.588794 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wbzsb" Oct 13 08:48:22 crc kubenswrapper[4685]: I1013 08:48:22.979840 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 08:48:22 crc kubenswrapper[4685]: I1013 08:48:22.980255 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 08:48:22 crc kubenswrapper[4685]: I1013 08:48:22.980325 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 08:48:22 crc kubenswrapper[4685]: I1013 08:48:22.981057 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008"} pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 08:48:22 crc kubenswrapper[4685]: I1013 08:48:22.981126 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" containerID="cri-o://d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008" gracePeriod=600 Oct 13 08:48:24 crc kubenswrapper[4685]: I1013 08:48:24.099244 4685 generic.go:334] "Generic (PLEG): container finished" podID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerID="d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008" exitCode=0 Oct 13 08:48:24 crc kubenswrapper[4685]: I1013 08:48:24.099344 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerDied","Data":"d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008"} Oct 13 08:48:24 crc kubenswrapper[4685]: I1013 08:48:24.099984 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerStarted","Data":"ab8bfde2e8129e04a0298b75eb5e60a218ebbe300f76aed267153fb073287fc0"} Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.347231 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" podUID="5072c69e-c292-478d-9998-8b5237dfc33f" containerName="oauth-openshift" containerID="cri-o://151f27404ab90dc5db9ba8ef80f161f8af4bca5ce732becc722a3fc05e7504c4" gracePeriod=15 Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.700098 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.735638 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-8f56ccf5-7pxp5"] Oct 13 08:48:30 crc kubenswrapper[4685]: E1013 08:48:30.735936 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c18864ff-4530-4655-b491-25f85b5604b4" containerName="extract-content" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.735951 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c18864ff-4530-4655-b491-25f85b5604b4" containerName="extract-content" Oct 13 08:48:30 crc kubenswrapper[4685]: E1013 08:48:30.735967 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c18864ff-4530-4655-b491-25f85b5604b4" containerName="extract-utilities" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.735975 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c18864ff-4530-4655-b491-25f85b5604b4" containerName="extract-utilities" Oct 13 08:48:30 crc kubenswrapper[4685]: E1013 08:48:30.735988 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a9fb57f-9b71-4922-aa37-2661e199871d" containerName="extract-content" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.735996 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a9fb57f-9b71-4922-aa37-2661e199871d" containerName="extract-content" Oct 13 08:48:30 crc kubenswrapper[4685]: E1013 08:48:30.736009 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a9fb57f-9b71-4922-aa37-2661e199871d" containerName="extract-utilities" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.736016 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a9fb57f-9b71-4922-aa37-2661e199871d" containerName="extract-utilities" Oct 13 08:48:30 crc kubenswrapper[4685]: E1013 08:48:30.736028 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a006d49-7467-4c6a-9024-09de1da2d8da" containerName="extract-content" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.736034 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a006d49-7467-4c6a-9024-09de1da2d8da" containerName="extract-content" Oct 13 08:48:30 crc kubenswrapper[4685]: E1013 08:48:30.736044 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5072c69e-c292-478d-9998-8b5237dfc33f" containerName="oauth-openshift" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.736051 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="5072c69e-c292-478d-9998-8b5237dfc33f" containerName="oauth-openshift" Oct 13 08:48:30 crc kubenswrapper[4685]: E1013 08:48:30.736059 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a006d49-7467-4c6a-9024-09de1da2d8da" containerName="registry-server" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.736066 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a006d49-7467-4c6a-9024-09de1da2d8da" containerName="registry-server" Oct 13 08:48:30 crc kubenswrapper[4685]: E1013 08:48:30.736076 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a9fb57f-9b71-4922-aa37-2661e199871d" containerName="registry-server" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.736084 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a9fb57f-9b71-4922-aa37-2661e199871d" containerName="registry-server" Oct 13 08:48:30 crc kubenswrapper[4685]: E1013 08:48:30.736101 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8beb316d-efe3-427b-a377-c84063b2b5d6" containerName="registry-server" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.736108 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="8beb316d-efe3-427b-a377-c84063b2b5d6" containerName="registry-server" Oct 13 08:48:30 crc kubenswrapper[4685]: E1013 08:48:30.736118 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a006d49-7467-4c6a-9024-09de1da2d8da" containerName="extract-utilities" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.736125 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a006d49-7467-4c6a-9024-09de1da2d8da" containerName="extract-utilities" Oct 13 08:48:30 crc kubenswrapper[4685]: E1013 08:48:30.736137 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8beb316d-efe3-427b-a377-c84063b2b5d6" containerName="extract-utilities" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.736144 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="8beb316d-efe3-427b-a377-c84063b2b5d6" containerName="extract-utilities" Oct 13 08:48:30 crc kubenswrapper[4685]: E1013 08:48:30.736157 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acf8bfed-51a6-4317-a60e-8b8ff193f47d" containerName="pruner" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.736164 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="acf8bfed-51a6-4317-a60e-8b8ff193f47d" containerName="pruner" Oct 13 08:48:30 crc kubenswrapper[4685]: E1013 08:48:30.736172 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8beb316d-efe3-427b-a377-c84063b2b5d6" containerName="extract-content" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.736178 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="8beb316d-efe3-427b-a377-c84063b2b5d6" containerName="extract-content" Oct 13 08:48:30 crc kubenswrapper[4685]: E1013 08:48:30.736187 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c18864ff-4530-4655-b491-25f85b5604b4" containerName="registry-server" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.736194 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c18864ff-4530-4655-b491-25f85b5604b4" containerName="registry-server" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.736300 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="acf8bfed-51a6-4317-a60e-8b8ff193f47d" containerName="pruner" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.736311 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="8beb316d-efe3-427b-a377-c84063b2b5d6" containerName="registry-server" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.736321 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a006d49-7467-4c6a-9024-09de1da2d8da" containerName="registry-server" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.736334 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a9fb57f-9b71-4922-aa37-2661e199871d" containerName="registry-server" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.736346 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="5072c69e-c292-478d-9998-8b5237dfc33f" containerName="oauth-openshift" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.736354 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="c18864ff-4530-4655-b491-25f85b5604b4" containerName="registry-server" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.737117 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.753546 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-8f56ccf5-7pxp5"] Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.789253 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-ocp-branding-template\") pod \"5072c69e-c292-478d-9998-8b5237dfc33f\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.789320 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-template-login\") pod \"5072c69e-c292-478d-9998-8b5237dfc33f\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.789354 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-session\") pod \"5072c69e-c292-478d-9998-8b5237dfc33f\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.789402 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-template-error\") pod \"5072c69e-c292-478d-9998-8b5237dfc33f\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.789429 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-trusted-ca-bundle\") pod \"5072c69e-c292-478d-9998-8b5237dfc33f\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.789503 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-template-provider-selection\") pod \"5072c69e-c292-478d-9998-8b5237dfc33f\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.789532 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5072c69e-c292-478d-9998-8b5237dfc33f-audit-dir\") pod \"5072c69e-c292-478d-9998-8b5237dfc33f\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.789573 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5k2b5\" (UniqueName: \"kubernetes.io/projected/5072c69e-c292-478d-9998-8b5237dfc33f-kube-api-access-5k2b5\") pod \"5072c69e-c292-478d-9998-8b5237dfc33f\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.789597 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-cliconfig\") pod \"5072c69e-c292-478d-9998-8b5237dfc33f\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.789627 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-idp-0-file-data\") pod \"5072c69e-c292-478d-9998-8b5237dfc33f\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.789656 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-router-certs\") pod \"5072c69e-c292-478d-9998-8b5237dfc33f\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.789690 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-audit-policies\") pod \"5072c69e-c292-478d-9998-8b5237dfc33f\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.789767 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-serving-cert\") pod \"5072c69e-c292-478d-9998-8b5237dfc33f\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.789790 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-service-ca\") pod \"5072c69e-c292-478d-9998-8b5237dfc33f\" (UID: \"5072c69e-c292-478d-9998-8b5237dfc33f\") " Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.790000 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-system-session\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.790026 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.790055 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.790085 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.790105 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.790121 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.790143 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-system-router-certs\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.790171 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eefb70a7-1093-4ef1-b710-b73e09cea7e5-audit-policies\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.790208 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wmk4\" (UniqueName: \"kubernetes.io/projected/eefb70a7-1093-4ef1-b710-b73e09cea7e5-kube-api-access-4wmk4\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.790231 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eefb70a7-1093-4ef1-b710-b73e09cea7e5-audit-dir\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.790254 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-user-template-error\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.790284 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-system-service-ca\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.790305 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-user-template-login\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.790341 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.791417 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "5072c69e-c292-478d-9998-8b5237dfc33f" (UID: "5072c69e-c292-478d-9998-8b5237dfc33f"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.794521 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "5072c69e-c292-478d-9998-8b5237dfc33f" (UID: "5072c69e-c292-478d-9998-8b5237dfc33f"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.795019 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "5072c69e-c292-478d-9998-8b5237dfc33f" (UID: "5072c69e-c292-478d-9998-8b5237dfc33f"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.795269 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "5072c69e-c292-478d-9998-8b5237dfc33f" (UID: "5072c69e-c292-478d-9998-8b5237dfc33f"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.798391 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "5072c69e-c292-478d-9998-8b5237dfc33f" (UID: "5072c69e-c292-478d-9998-8b5237dfc33f"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.798616 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5072c69e-c292-478d-9998-8b5237dfc33f-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "5072c69e-c292-478d-9998-8b5237dfc33f" (UID: "5072c69e-c292-478d-9998-8b5237dfc33f"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.805139 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "5072c69e-c292-478d-9998-8b5237dfc33f" (UID: "5072c69e-c292-478d-9998-8b5237dfc33f"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.805159 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5072c69e-c292-478d-9998-8b5237dfc33f-kube-api-access-5k2b5" (OuterVolumeSpecName: "kube-api-access-5k2b5") pod "5072c69e-c292-478d-9998-8b5237dfc33f" (UID: "5072c69e-c292-478d-9998-8b5237dfc33f"). InnerVolumeSpecName "kube-api-access-5k2b5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.805947 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "5072c69e-c292-478d-9998-8b5237dfc33f" (UID: "5072c69e-c292-478d-9998-8b5237dfc33f"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.806383 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "5072c69e-c292-478d-9998-8b5237dfc33f" (UID: "5072c69e-c292-478d-9998-8b5237dfc33f"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.807318 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "5072c69e-c292-478d-9998-8b5237dfc33f" (UID: "5072c69e-c292-478d-9998-8b5237dfc33f"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.807758 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "5072c69e-c292-478d-9998-8b5237dfc33f" (UID: "5072c69e-c292-478d-9998-8b5237dfc33f"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.808629 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "5072c69e-c292-478d-9998-8b5237dfc33f" (UID: "5072c69e-c292-478d-9998-8b5237dfc33f"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.811358 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "5072c69e-c292-478d-9998-8b5237dfc33f" (UID: "5072c69e-c292-478d-9998-8b5237dfc33f"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.891558 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eefb70a7-1093-4ef1-b710-b73e09cea7e5-audit-policies\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.891626 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wmk4\" (UniqueName: \"kubernetes.io/projected/eefb70a7-1093-4ef1-b710-b73e09cea7e5-kube-api-access-4wmk4\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.891650 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eefb70a7-1093-4ef1-b710-b73e09cea7e5-audit-dir\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.891666 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-user-template-error\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.891693 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-system-service-ca\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.891714 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-user-template-login\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.891738 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.891758 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-system-session\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.891777 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.891800 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.891826 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.891851 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.891874 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.891899 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-system-router-certs\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.892003 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.892044 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.892056 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.892071 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.892629 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eefb70a7-1093-4ef1-b710-b73e09cea7e5-audit-policies\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.893372 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.894085 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.894107 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.894119 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.894131 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.894160 4685 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5072c69e-c292-478d-9998-8b5237dfc33f-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.894174 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5k2b5\" (UniqueName: \"kubernetes.io/projected/5072c69e-c292-478d-9998-8b5237dfc33f-kube-api-access-5k2b5\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.894186 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.894198 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.894210 4685 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5072c69e-c292-478d-9998-8b5237dfc33f-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.894242 4685 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5072c69e-c292-478d-9998-8b5237dfc33f-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.896102 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.896355 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.896364 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eefb70a7-1093-4ef1-b710-b73e09cea7e5-audit-dir\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.896456 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.896507 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-system-service-ca\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.897060 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-user-template-login\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.897379 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.898004 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-system-router-certs\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.898364 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-system-session\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.898589 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-user-template-error\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.899498 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/eefb70a7-1093-4ef1-b710-b73e09cea7e5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:30 crc kubenswrapper[4685]: I1013 08:48:30.907871 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wmk4\" (UniqueName: \"kubernetes.io/projected/eefb70a7-1093-4ef1-b710-b73e09cea7e5-kube-api-access-4wmk4\") pod \"oauth-openshift-8f56ccf5-7pxp5\" (UID: \"eefb70a7-1093-4ef1-b710-b73e09cea7e5\") " pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:31 crc kubenswrapper[4685]: I1013 08:48:31.063806 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:31 crc kubenswrapper[4685]: I1013 08:48:31.141154 4685 generic.go:334] "Generic (PLEG): container finished" podID="5072c69e-c292-478d-9998-8b5237dfc33f" containerID="151f27404ab90dc5db9ba8ef80f161f8af4bca5ce732becc722a3fc05e7504c4" exitCode=0 Oct 13 08:48:31 crc kubenswrapper[4685]: I1013 08:48:31.141201 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" event={"ID":"5072c69e-c292-478d-9998-8b5237dfc33f","Type":"ContainerDied","Data":"151f27404ab90dc5db9ba8ef80f161f8af4bca5ce732becc722a3fc05e7504c4"} Oct 13 08:48:31 crc kubenswrapper[4685]: I1013 08:48:31.141232 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" event={"ID":"5072c69e-c292-478d-9998-8b5237dfc33f","Type":"ContainerDied","Data":"ce543441c286144011a8d80fdbf12acb00250210fa9082c11b50b6d1ff64be0d"} Oct 13 08:48:31 crc kubenswrapper[4685]: I1013 08:48:31.141252 4685 scope.go:117] "RemoveContainer" containerID="151f27404ab90dc5db9ba8ef80f161f8af4bca5ce732becc722a3fc05e7504c4" Oct 13 08:48:31 crc kubenswrapper[4685]: I1013 08:48:31.141392 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-54dh8" Oct 13 08:48:31 crc kubenswrapper[4685]: I1013 08:48:31.168113 4685 scope.go:117] "RemoveContainer" containerID="151f27404ab90dc5db9ba8ef80f161f8af4bca5ce732becc722a3fc05e7504c4" Oct 13 08:48:31 crc kubenswrapper[4685]: I1013 08:48:31.179599 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-54dh8"] Oct 13 08:48:31 crc kubenswrapper[4685]: E1013 08:48:31.182636 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"151f27404ab90dc5db9ba8ef80f161f8af4bca5ce732becc722a3fc05e7504c4\": container with ID starting with 151f27404ab90dc5db9ba8ef80f161f8af4bca5ce732becc722a3fc05e7504c4 not found: ID does not exist" containerID="151f27404ab90dc5db9ba8ef80f161f8af4bca5ce732becc722a3fc05e7504c4" Oct 13 08:48:31 crc kubenswrapper[4685]: I1013 08:48:31.182686 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"151f27404ab90dc5db9ba8ef80f161f8af4bca5ce732becc722a3fc05e7504c4"} err="failed to get container status \"151f27404ab90dc5db9ba8ef80f161f8af4bca5ce732becc722a3fc05e7504c4\": rpc error: code = NotFound desc = could not find container \"151f27404ab90dc5db9ba8ef80f161f8af4bca5ce732becc722a3fc05e7504c4\": container with ID starting with 151f27404ab90dc5db9ba8ef80f161f8af4bca5ce732becc722a3fc05e7504c4 not found: ID does not exist" Oct 13 08:48:31 crc kubenswrapper[4685]: I1013 08:48:31.189797 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-54dh8"] Oct 13 08:48:31 crc kubenswrapper[4685]: I1013 08:48:31.512698 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5072c69e-c292-478d-9998-8b5237dfc33f" path="/var/lib/kubelet/pods/5072c69e-c292-478d-9998-8b5237dfc33f/volumes" Oct 13 08:48:31 crc kubenswrapper[4685]: I1013 08:48:31.534056 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-8f56ccf5-7pxp5"] Oct 13 08:48:32 crc kubenswrapper[4685]: I1013 08:48:32.148195 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" event={"ID":"eefb70a7-1093-4ef1-b710-b73e09cea7e5","Type":"ContainerStarted","Data":"26449b8aa3b4852ffefbb63a50253a3c6f14ab917beeb0c539dd56d83ec76554"} Oct 13 08:48:32 crc kubenswrapper[4685]: I1013 08:48:32.149169 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" event={"ID":"eefb70a7-1093-4ef1-b710-b73e09cea7e5","Type":"ContainerStarted","Data":"776d24d905b8474c56794aa800e7fcd83a89d895d10e11082a3aa7f506660d3c"} Oct 13 08:48:32 crc kubenswrapper[4685]: I1013 08:48:32.149199 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:48:32 crc kubenswrapper[4685]: I1013 08:48:32.167769 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" podStartSLOduration=27.167746699 podStartE2EDuration="27.167746699s" podCreationTimestamp="2025-10-13 08:48:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:48:32.16660184 +0000 UTC m=+237.314477601" watchObservedRunningTime="2025-10-13 08:48:32.167746699 +0000 UTC m=+237.315622460" Oct 13 08:48:32 crc kubenswrapper[4685]: I1013 08:48:32.175008 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-8f56ccf5-7pxp5" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.085299 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wbzsb"] Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.088562 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wbzsb" podUID="1ca80dd4-6294-4263-b938-188c17814556" containerName="registry-server" containerID="cri-o://f001baad0e6618d53b2401c1e76f5498a24e3b3c4fa88b92f6b681279337490f" gracePeriod=30 Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.098457 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vdbjj"] Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.098830 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vdbjj" podUID="23545a99-b853-4a78-85a6-6186dc3bf903" containerName="registry-server" containerID="cri-o://6941be18324ed1a356f9c714f087aad7dd4be8224e0c1fc1ba27f4a5d5c846f7" gracePeriod=30 Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.107022 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7sx5f"] Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.107294 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" podUID="b8ae1ce1-bf5f-4d13-87b4-0810f63434af" containerName="marketplace-operator" containerID="cri-o://6379ee3dd8ba0c8279c34e7d87be880c84fb191bd775ff824985be478e72968d" gracePeriod=30 Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.119264 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fmmfz"] Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.119644 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fmmfz" podUID="e74b537e-de07-4a47-85e7-e2bd5a72a11c" containerName="registry-server" containerID="cri-o://c743e565b6746e20610f8f71c8279e78db342f92df329e8b719fa201fafe481c" gracePeriod=30 Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.127820 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v6zv9"] Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.128496 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-v6zv9" podUID="75521184-74d3-44b9-a410-b0b0fafa4384" containerName="registry-server" containerID="cri-o://14cd7f4d36a1d55aadc343b5a23b83f99053aa71adbfaf0b71f35ba7ccc783c9" gracePeriod=30 Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.147346 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5pssn"] Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.148217 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5pssn" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.168693 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5pssn"] Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.242450 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/30028087-7eef-4beb-8011-304f2dce2e01-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5pssn\" (UID: \"30028087-7eef-4beb-8011-304f2dce2e01\") " pod="openshift-marketplace/marketplace-operator-79b997595-5pssn" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.242569 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxhw2\" (UniqueName: \"kubernetes.io/projected/30028087-7eef-4beb-8011-304f2dce2e01-kube-api-access-xxhw2\") pod \"marketplace-operator-79b997595-5pssn\" (UID: \"30028087-7eef-4beb-8011-304f2dce2e01\") " pod="openshift-marketplace/marketplace-operator-79b997595-5pssn" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.242635 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/30028087-7eef-4beb-8011-304f2dce2e01-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5pssn\" (UID: \"30028087-7eef-4beb-8011-304f2dce2e01\") " pod="openshift-marketplace/marketplace-operator-79b997595-5pssn" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.344952 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxhw2\" (UniqueName: \"kubernetes.io/projected/30028087-7eef-4beb-8011-304f2dce2e01-kube-api-access-xxhw2\") pod \"marketplace-operator-79b997595-5pssn\" (UID: \"30028087-7eef-4beb-8011-304f2dce2e01\") " pod="openshift-marketplace/marketplace-operator-79b997595-5pssn" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.345346 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/30028087-7eef-4beb-8011-304f2dce2e01-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5pssn\" (UID: \"30028087-7eef-4beb-8011-304f2dce2e01\") " pod="openshift-marketplace/marketplace-operator-79b997595-5pssn" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.345381 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/30028087-7eef-4beb-8011-304f2dce2e01-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5pssn\" (UID: \"30028087-7eef-4beb-8011-304f2dce2e01\") " pod="openshift-marketplace/marketplace-operator-79b997595-5pssn" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.346849 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/30028087-7eef-4beb-8011-304f2dce2e01-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5pssn\" (UID: \"30028087-7eef-4beb-8011-304f2dce2e01\") " pod="openshift-marketplace/marketplace-operator-79b997595-5pssn" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.353611 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/30028087-7eef-4beb-8011-304f2dce2e01-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5pssn\" (UID: \"30028087-7eef-4beb-8011-304f2dce2e01\") " pod="openshift-marketplace/marketplace-operator-79b997595-5pssn" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.371992 4685 generic.go:334] "Generic (PLEG): container finished" podID="75521184-74d3-44b9-a410-b0b0fafa4384" containerID="14cd7f4d36a1d55aadc343b5a23b83f99053aa71adbfaf0b71f35ba7ccc783c9" exitCode=0 Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.372125 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6zv9" event={"ID":"75521184-74d3-44b9-a410-b0b0fafa4384","Type":"ContainerDied","Data":"14cd7f4d36a1d55aadc343b5a23b83f99053aa71adbfaf0b71f35ba7ccc783c9"} Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.374387 4685 generic.go:334] "Generic (PLEG): container finished" podID="b8ae1ce1-bf5f-4d13-87b4-0810f63434af" containerID="6379ee3dd8ba0c8279c34e7d87be880c84fb191bd775ff824985be478e72968d" exitCode=0 Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.374454 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" event={"ID":"b8ae1ce1-bf5f-4d13-87b4-0810f63434af","Type":"ContainerDied","Data":"6379ee3dd8ba0c8279c34e7d87be880c84fb191bd775ff824985be478e72968d"} Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.375749 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxhw2\" (UniqueName: \"kubernetes.io/projected/30028087-7eef-4beb-8011-304f2dce2e01-kube-api-access-xxhw2\") pod \"marketplace-operator-79b997595-5pssn\" (UID: \"30028087-7eef-4beb-8011-304f2dce2e01\") " pod="openshift-marketplace/marketplace-operator-79b997595-5pssn" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.378357 4685 generic.go:334] "Generic (PLEG): container finished" podID="1ca80dd4-6294-4263-b938-188c17814556" containerID="f001baad0e6618d53b2401c1e76f5498a24e3b3c4fa88b92f6b681279337490f" exitCode=0 Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.378404 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbzsb" event={"ID":"1ca80dd4-6294-4263-b938-188c17814556","Type":"ContainerDied","Data":"f001baad0e6618d53b2401c1e76f5498a24e3b3c4fa88b92f6b681279337490f"} Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.385126 4685 generic.go:334] "Generic (PLEG): container finished" podID="23545a99-b853-4a78-85a6-6186dc3bf903" containerID="6941be18324ed1a356f9c714f087aad7dd4be8224e0c1fc1ba27f4a5d5c846f7" exitCode=0 Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.385208 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdbjj" event={"ID":"23545a99-b853-4a78-85a6-6186dc3bf903","Type":"ContainerDied","Data":"6941be18324ed1a356f9c714f087aad7dd4be8224e0c1fc1ba27f4a5d5c846f7"} Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.390792 4685 generic.go:334] "Generic (PLEG): container finished" podID="e74b537e-de07-4a47-85e7-e2bd5a72a11c" containerID="c743e565b6746e20610f8f71c8279e78db342f92df329e8b719fa201fafe481c" exitCode=0 Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.390887 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fmmfz" event={"ID":"e74b537e-de07-4a47-85e7-e2bd5a72a11c","Type":"ContainerDied","Data":"c743e565b6746e20610f8f71c8279e78db342f92df329e8b719fa201fafe481c"} Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.472725 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5pssn" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.510488 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wbzsb" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.547147 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ca80dd4-6294-4263-b938-188c17814556-catalog-content\") pod \"1ca80dd4-6294-4263-b938-188c17814556\" (UID: \"1ca80dd4-6294-4263-b938-188c17814556\") " Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.547186 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rf74x\" (UniqueName: \"kubernetes.io/projected/1ca80dd4-6294-4263-b938-188c17814556-kube-api-access-rf74x\") pod \"1ca80dd4-6294-4263-b938-188c17814556\" (UID: \"1ca80dd4-6294-4263-b938-188c17814556\") " Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.547213 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ca80dd4-6294-4263-b938-188c17814556-utilities\") pod \"1ca80dd4-6294-4263-b938-188c17814556\" (UID: \"1ca80dd4-6294-4263-b938-188c17814556\") " Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.550247 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ca80dd4-6294-4263-b938-188c17814556-utilities" (OuterVolumeSpecName: "utilities") pod "1ca80dd4-6294-4263-b938-188c17814556" (UID: "1ca80dd4-6294-4263-b938-188c17814556"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:49:07 crc kubenswrapper[4685]: E1013 08:49:07.552980 4685 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c743e565b6746e20610f8f71c8279e78db342f92df329e8b719fa201fafe481c is running failed: container process not found" containerID="c743e565b6746e20610f8f71c8279e78db342f92df329e8b719fa201fafe481c" cmd=["grpc_health_probe","-addr=:50051"] Oct 13 08:49:07 crc kubenswrapper[4685]: E1013 08:49:07.553264 4685 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c743e565b6746e20610f8f71c8279e78db342f92df329e8b719fa201fafe481c is running failed: container process not found" containerID="c743e565b6746e20610f8f71c8279e78db342f92df329e8b719fa201fafe481c" cmd=["grpc_health_probe","-addr=:50051"] Oct 13 08:49:07 crc kubenswrapper[4685]: E1013 08:49:07.553850 4685 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c743e565b6746e20610f8f71c8279e78db342f92df329e8b719fa201fafe481c is running failed: container process not found" containerID="c743e565b6746e20610f8f71c8279e78db342f92df329e8b719fa201fafe481c" cmd=["grpc_health_probe","-addr=:50051"] Oct 13 08:49:07 crc kubenswrapper[4685]: E1013 08:49:07.553885 4685 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c743e565b6746e20610f8f71c8279e78db342f92df329e8b719fa201fafe481c is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-fmmfz" podUID="e74b537e-de07-4a47-85e7-e2bd5a72a11c" containerName="registry-server" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.557175 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ca80dd4-6294-4263-b938-188c17814556-kube-api-access-rf74x" (OuterVolumeSpecName: "kube-api-access-rf74x") pod "1ca80dd4-6294-4263-b938-188c17814556" (UID: "1ca80dd4-6294-4263-b938-188c17814556"). InnerVolumeSpecName "kube-api-access-rf74x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.637879 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ca80dd4-6294-4263-b938-188c17814556-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ca80dd4-6294-4263-b938-188c17814556" (UID: "1ca80dd4-6294-4263-b938-188c17814556"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.650833 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ca80dd4-6294-4263-b938-188c17814556-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.652354 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rf74x\" (UniqueName: \"kubernetes.io/projected/1ca80dd4-6294-4263-b938-188c17814556-kube-api-access-rf74x\") on node \"crc\" DevicePath \"\"" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.661841 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ca80dd4-6294-4263-b938-188c17814556-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.716530 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vdbjj" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.736070 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.770871 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b8ae1ce1-bf5f-4d13-87b4-0810f63434af-marketplace-operator-metrics\") pod \"b8ae1ce1-bf5f-4d13-87b4-0810f63434af\" (UID: \"b8ae1ce1-bf5f-4d13-87b4-0810f63434af\") " Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.770963 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23545a99-b853-4a78-85a6-6186dc3bf903-catalog-content\") pod \"23545a99-b853-4a78-85a6-6186dc3bf903\" (UID: \"23545a99-b853-4a78-85a6-6186dc3bf903\") " Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.771747 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zbqc\" (UniqueName: \"kubernetes.io/projected/b8ae1ce1-bf5f-4d13-87b4-0810f63434af-kube-api-access-9zbqc\") pod \"b8ae1ce1-bf5f-4d13-87b4-0810f63434af\" (UID: \"b8ae1ce1-bf5f-4d13-87b4-0810f63434af\") " Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.771873 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b8ae1ce1-bf5f-4d13-87b4-0810f63434af-marketplace-trusted-ca\") pod \"b8ae1ce1-bf5f-4d13-87b4-0810f63434af\" (UID: \"b8ae1ce1-bf5f-4d13-87b4-0810f63434af\") " Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.772039 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ljd6\" (UniqueName: \"kubernetes.io/projected/23545a99-b853-4a78-85a6-6186dc3bf903-kube-api-access-8ljd6\") pod \"23545a99-b853-4a78-85a6-6186dc3bf903\" (UID: \"23545a99-b853-4a78-85a6-6186dc3bf903\") " Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.772071 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23545a99-b853-4a78-85a6-6186dc3bf903-utilities\") pod \"23545a99-b853-4a78-85a6-6186dc3bf903\" (UID: \"23545a99-b853-4a78-85a6-6186dc3bf903\") " Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.773273 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23545a99-b853-4a78-85a6-6186dc3bf903-utilities" (OuterVolumeSpecName: "utilities") pod "23545a99-b853-4a78-85a6-6186dc3bf903" (UID: "23545a99-b853-4a78-85a6-6186dc3bf903"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.775973 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8ae1ce1-bf5f-4d13-87b4-0810f63434af-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b8ae1ce1-bf5f-4d13-87b4-0810f63434af" (UID: "b8ae1ce1-bf5f-4d13-87b4-0810f63434af"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.787035 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8ae1ce1-bf5f-4d13-87b4-0810f63434af-kube-api-access-9zbqc" (OuterVolumeSpecName: "kube-api-access-9zbqc") pod "b8ae1ce1-bf5f-4d13-87b4-0810f63434af" (UID: "b8ae1ce1-bf5f-4d13-87b4-0810f63434af"). InnerVolumeSpecName "kube-api-access-9zbqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.787283 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8ae1ce1-bf5f-4d13-87b4-0810f63434af-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b8ae1ce1-bf5f-4d13-87b4-0810f63434af" (UID: "b8ae1ce1-bf5f-4d13-87b4-0810f63434af"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.802265 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23545a99-b853-4a78-85a6-6186dc3bf903-kube-api-access-8ljd6" (OuterVolumeSpecName: "kube-api-access-8ljd6") pod "23545a99-b853-4a78-85a6-6186dc3bf903" (UID: "23545a99-b853-4a78-85a6-6186dc3bf903"). InnerVolumeSpecName "kube-api-access-8ljd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.834808 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fmmfz" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.842089 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v6zv9" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.868149 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23545a99-b853-4a78-85a6-6186dc3bf903-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "23545a99-b853-4a78-85a6-6186dc3bf903" (UID: "23545a99-b853-4a78-85a6-6186dc3bf903"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.873515 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75521184-74d3-44b9-a410-b0b0fafa4384-catalog-content\") pod \"75521184-74d3-44b9-a410-b0b0fafa4384\" (UID: \"75521184-74d3-44b9-a410-b0b0fafa4384\") " Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.873580 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75521184-74d3-44b9-a410-b0b0fafa4384-utilities\") pod \"75521184-74d3-44b9-a410-b0b0fafa4384\" (UID: \"75521184-74d3-44b9-a410-b0b0fafa4384\") " Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.873610 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drl5f\" (UniqueName: \"kubernetes.io/projected/75521184-74d3-44b9-a410-b0b0fafa4384-kube-api-access-drl5f\") pod \"75521184-74d3-44b9-a410-b0b0fafa4384\" (UID: \"75521184-74d3-44b9-a410-b0b0fafa4384\") " Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.873671 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e74b537e-de07-4a47-85e7-e2bd5a72a11c-catalog-content\") pod \"e74b537e-de07-4a47-85e7-e2bd5a72a11c\" (UID: \"e74b537e-de07-4a47-85e7-e2bd5a72a11c\") " Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.873714 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e74b537e-de07-4a47-85e7-e2bd5a72a11c-utilities\") pod \"e74b537e-de07-4a47-85e7-e2bd5a72a11c\" (UID: \"e74b537e-de07-4a47-85e7-e2bd5a72a11c\") " Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.873770 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrcfq\" (UniqueName: \"kubernetes.io/projected/e74b537e-de07-4a47-85e7-e2bd5a72a11c-kube-api-access-rrcfq\") pod \"e74b537e-de07-4a47-85e7-e2bd5a72a11c\" (UID: \"e74b537e-de07-4a47-85e7-e2bd5a72a11c\") " Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.874219 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zbqc\" (UniqueName: \"kubernetes.io/projected/b8ae1ce1-bf5f-4d13-87b4-0810f63434af-kube-api-access-9zbqc\") on node \"crc\" DevicePath \"\"" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.874236 4685 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b8ae1ce1-bf5f-4d13-87b4-0810f63434af-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.874248 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ljd6\" (UniqueName: \"kubernetes.io/projected/23545a99-b853-4a78-85a6-6186dc3bf903-kube-api-access-8ljd6\") on node \"crc\" DevicePath \"\"" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.874261 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23545a99-b853-4a78-85a6-6186dc3bf903-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.874273 4685 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b8ae1ce1-bf5f-4d13-87b4-0810f63434af-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.874283 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23545a99-b853-4a78-85a6-6186dc3bf903-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.877350 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e74b537e-de07-4a47-85e7-e2bd5a72a11c-utilities" (OuterVolumeSpecName: "utilities") pod "e74b537e-de07-4a47-85e7-e2bd5a72a11c" (UID: "e74b537e-de07-4a47-85e7-e2bd5a72a11c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.878148 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75521184-74d3-44b9-a410-b0b0fafa4384-utilities" (OuterVolumeSpecName: "utilities") pod "75521184-74d3-44b9-a410-b0b0fafa4384" (UID: "75521184-74d3-44b9-a410-b0b0fafa4384"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.895382 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e74b537e-de07-4a47-85e7-e2bd5a72a11c-kube-api-access-rrcfq" (OuterVolumeSpecName: "kube-api-access-rrcfq") pod "e74b537e-de07-4a47-85e7-e2bd5a72a11c" (UID: "e74b537e-de07-4a47-85e7-e2bd5a72a11c"). InnerVolumeSpecName "kube-api-access-rrcfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.902257 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75521184-74d3-44b9-a410-b0b0fafa4384-kube-api-access-drl5f" (OuterVolumeSpecName: "kube-api-access-drl5f") pod "75521184-74d3-44b9-a410-b0b0fafa4384" (UID: "75521184-74d3-44b9-a410-b0b0fafa4384"). InnerVolumeSpecName "kube-api-access-drl5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.932616 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e74b537e-de07-4a47-85e7-e2bd5a72a11c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e74b537e-de07-4a47-85e7-e2bd5a72a11c" (UID: "e74b537e-de07-4a47-85e7-e2bd5a72a11c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.952626 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5pssn"] Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.975471 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75521184-74d3-44b9-a410-b0b0fafa4384-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.975505 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drl5f\" (UniqueName: \"kubernetes.io/projected/75521184-74d3-44b9-a410-b0b0fafa4384-kube-api-access-drl5f\") on node \"crc\" DevicePath \"\"" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.975517 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e74b537e-de07-4a47-85e7-e2bd5a72a11c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.975527 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e74b537e-de07-4a47-85e7-e2bd5a72a11c-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.975536 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrcfq\" (UniqueName: \"kubernetes.io/projected/e74b537e-de07-4a47-85e7-e2bd5a72a11c-kube-api-access-rrcfq\") on node \"crc\" DevicePath \"\"" Oct 13 08:49:07 crc kubenswrapper[4685]: I1013 08:49:07.977318 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75521184-74d3-44b9-a410-b0b0fafa4384-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "75521184-74d3-44b9-a410-b0b0fafa4384" (UID: "75521184-74d3-44b9-a410-b0b0fafa4384"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.077158 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75521184-74d3-44b9-a410-b0b0fafa4384-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.397648 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5pssn" event={"ID":"30028087-7eef-4beb-8011-304f2dce2e01","Type":"ContainerStarted","Data":"5d17e82fe50aa28e2f664da13f2dccaefceb50fb20cc096064a0ee577aef5c12"} Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.398337 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5pssn" event={"ID":"30028087-7eef-4beb-8011-304f2dce2e01","Type":"ContainerStarted","Data":"e6e9f992e92a3f98c934ff92554e1d438fd012630b82b353df65ba45683f554a"} Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.398398 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5pssn" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.402314 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6zv9" event={"ID":"75521184-74d3-44b9-a410-b0b0fafa4384","Type":"ContainerDied","Data":"f53264f68a3d9bc80305a3b0bf45176413abd2e224dccf6c8f282f9b6ed05522"} Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.402357 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v6zv9" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.402391 4685 scope.go:117] "RemoveContainer" containerID="14cd7f4d36a1d55aadc343b5a23b83f99053aa71adbfaf0b71f35ba7ccc783c9" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.402864 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-5pssn" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.404874 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" event={"ID":"b8ae1ce1-bf5f-4d13-87b4-0810f63434af","Type":"ContainerDied","Data":"e961755b572021234363fccf249332fc4b3c12ce669dfb40dba16f9676a3689e"} Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.404905 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7sx5f" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.420362 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbzsb" event={"ID":"1ca80dd4-6294-4263-b938-188c17814556","Type":"ContainerDied","Data":"1f02035b8ba6c426061e6d3734c23df93ad75ccd361dae247d47f4f28f0c79a6"} Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.420484 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wbzsb" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.429269 4685 scope.go:117] "RemoveContainer" containerID="dcefc1b2b3e483363f52e5e450854eec214d122c593dcd06649fb752bb8bca77" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.435109 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-5pssn" podStartSLOduration=1.435075524 podStartE2EDuration="1.435075524s" podCreationTimestamp="2025-10-13 08:49:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:49:08.430522333 +0000 UTC m=+273.578398114" watchObservedRunningTime="2025-10-13 08:49:08.435075524 +0000 UTC m=+273.582951285" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.439510 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vdbjj" event={"ID":"23545a99-b853-4a78-85a6-6186dc3bf903","Type":"ContainerDied","Data":"c36885e110b59fb5a1bc436423fc8ca80824ee5b58e3d39ebb675a514d2496f9"} Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.439676 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vdbjj" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.460593 4685 scope.go:117] "RemoveContainer" containerID="71df0cf490c8514d2a1280ace7996d87626606dbfcc11bea3502cc3fc2386186" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.460724 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fmmfz" event={"ID":"e74b537e-de07-4a47-85e7-e2bd5a72a11c","Type":"ContainerDied","Data":"75a9741253f9702f47ad595be38cbfb122c4db6ec6b6ea1ec0077314f161a354"} Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.460855 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fmmfz" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.513264 4685 scope.go:117] "RemoveContainer" containerID="6379ee3dd8ba0c8279c34e7d87be880c84fb191bd775ff824985be478e72968d" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.535524 4685 scope.go:117] "RemoveContainer" containerID="f001baad0e6618d53b2401c1e76f5498a24e3b3c4fa88b92f6b681279337490f" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.550263 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7sx5f"] Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.554829 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7sx5f"] Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.558144 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v6zv9"] Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.560017 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-v6zv9"] Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.576149 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wbzsb"] Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.580423 4685 scope.go:117] "RemoveContainer" containerID="6942d52617cb26bd17e1bd92a76ce68ff8e2c5cb68cdefb33f1e0da6d43f8963" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.585228 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wbzsb"] Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.589066 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vdbjj"] Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.594200 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vdbjj"] Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.601145 4685 scope.go:117] "RemoveContainer" containerID="a355a3c958302c1e868888905c2e735fa87447330184e854d0c34cbd9627f830" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.607785 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fmmfz"] Oct 13 08:49:08 crc kubenswrapper[4685]: E1013 08:49:08.609228 4685 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75521184_74d3_44b9_a410_b0b0fafa4384.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ca80dd4_6294_4263_b938_188c17814556.slice/crio-1f02035b8ba6c426061e6d3734c23df93ad75ccd361dae247d47f4f28f0c79a6\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8ae1ce1_bf5f_4d13_87b4_0810f63434af.slice/crio-e961755b572021234363fccf249332fc4b3c12ce669dfb40dba16f9676a3689e\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23545a99_b853_4a78_85a6_6186dc3bf903.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode74b537e_de07_4a47_85e7_e2bd5a72a11c.slice\": RecentStats: unable to find data in memory cache]" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.611030 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fmmfz"] Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.631808 4685 scope.go:117] "RemoveContainer" containerID="6941be18324ed1a356f9c714f087aad7dd4be8224e0c1fc1ba27f4a5d5c846f7" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.647205 4685 scope.go:117] "RemoveContainer" containerID="4e07cceaa718b1733c04d9dca707713e6914cbc8e8d84f555b853f0ce507af46" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.663002 4685 scope.go:117] "RemoveContainer" containerID="b4fd611ac286faa6b033f5fbf108e793a8af2ebf53102ce520ec2303d8cddc36" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.681151 4685 scope.go:117] "RemoveContainer" containerID="c743e565b6746e20610f8f71c8279e78db342f92df329e8b719fa201fafe481c" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.705005 4685 scope.go:117] "RemoveContainer" containerID="3c437e73337675f69ec91870d24f037a180a0a577b56a2a3e093d134d92b6643" Oct 13 08:49:08 crc kubenswrapper[4685]: I1013 08:49:08.720027 4685 scope.go:117] "RemoveContainer" containerID="87412e9e92501e922f8d5586ec8004940e405e413e2494f1859a71fc3d072cce" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.308017 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t9v8k"] Oct 13 08:49:09 crc kubenswrapper[4685]: E1013 08:49:09.309707 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca80dd4-6294-4263-b938-188c17814556" containerName="extract-content" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.309883 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca80dd4-6294-4263-b938-188c17814556" containerName="extract-content" Oct 13 08:49:09 crc kubenswrapper[4685]: E1013 08:49:09.310055 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca80dd4-6294-4263-b938-188c17814556" containerName="registry-server" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.310199 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca80dd4-6294-4263-b938-188c17814556" containerName="registry-server" Oct 13 08:49:09 crc kubenswrapper[4685]: E1013 08:49:09.310297 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8ae1ce1-bf5f-4d13-87b4-0810f63434af" containerName="marketplace-operator" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.310392 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8ae1ce1-bf5f-4d13-87b4-0810f63434af" containerName="marketplace-operator" Oct 13 08:49:09 crc kubenswrapper[4685]: E1013 08:49:09.310529 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e74b537e-de07-4a47-85e7-e2bd5a72a11c" containerName="extract-content" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.310665 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e74b537e-de07-4a47-85e7-e2bd5a72a11c" containerName="extract-content" Oct 13 08:49:09 crc kubenswrapper[4685]: E1013 08:49:09.310774 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23545a99-b853-4a78-85a6-6186dc3bf903" containerName="extract-utilities" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.310874 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="23545a99-b853-4a78-85a6-6186dc3bf903" containerName="extract-utilities" Oct 13 08:49:09 crc kubenswrapper[4685]: E1013 08:49:09.310998 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca80dd4-6294-4263-b938-188c17814556" containerName="extract-utilities" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.311086 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca80dd4-6294-4263-b938-188c17814556" containerName="extract-utilities" Oct 13 08:49:09 crc kubenswrapper[4685]: E1013 08:49:09.311178 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23545a99-b853-4a78-85a6-6186dc3bf903" containerName="registry-server" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.311261 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="23545a99-b853-4a78-85a6-6186dc3bf903" containerName="registry-server" Oct 13 08:49:09 crc kubenswrapper[4685]: E1013 08:49:09.311364 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23545a99-b853-4a78-85a6-6186dc3bf903" containerName="extract-content" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.311452 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="23545a99-b853-4a78-85a6-6186dc3bf903" containerName="extract-content" Oct 13 08:49:09 crc kubenswrapper[4685]: E1013 08:49:09.311558 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e74b537e-de07-4a47-85e7-e2bd5a72a11c" containerName="registry-server" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.311655 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e74b537e-de07-4a47-85e7-e2bd5a72a11c" containerName="registry-server" Oct 13 08:49:09 crc kubenswrapper[4685]: E1013 08:49:09.311760 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75521184-74d3-44b9-a410-b0b0fafa4384" containerName="registry-server" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.312053 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="75521184-74d3-44b9-a410-b0b0fafa4384" containerName="registry-server" Oct 13 08:49:09 crc kubenswrapper[4685]: E1013 08:49:09.312157 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75521184-74d3-44b9-a410-b0b0fafa4384" containerName="extract-utilities" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.312258 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="75521184-74d3-44b9-a410-b0b0fafa4384" containerName="extract-utilities" Oct 13 08:49:09 crc kubenswrapper[4685]: E1013 08:49:09.312410 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75521184-74d3-44b9-a410-b0b0fafa4384" containerName="extract-content" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.312584 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="75521184-74d3-44b9-a410-b0b0fafa4384" containerName="extract-content" Oct 13 08:49:09 crc kubenswrapper[4685]: E1013 08:49:09.312792 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e74b537e-de07-4a47-85e7-e2bd5a72a11c" containerName="extract-utilities" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.312993 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e74b537e-de07-4a47-85e7-e2bd5a72a11c" containerName="extract-utilities" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.313282 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="23545a99-b853-4a78-85a6-6186dc3bf903" containerName="registry-server" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.313700 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="e74b537e-de07-4a47-85e7-e2bd5a72a11c" containerName="registry-server" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.313982 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca80dd4-6294-4263-b938-188c17814556" containerName="registry-server" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.314296 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="75521184-74d3-44b9-a410-b0b0fafa4384" containerName="registry-server" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.314472 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8ae1ce1-bf5f-4d13-87b4-0810f63434af" containerName="marketplace-operator" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.317543 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t9v8k" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.320430 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.327810 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t9v8k"] Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.419150 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9adb7fcf-2d97-4c07-85cf-86cef286dc69-utilities\") pod \"redhat-operators-t9v8k\" (UID: \"9adb7fcf-2d97-4c07-85cf-86cef286dc69\") " pod="openshift-marketplace/redhat-operators-t9v8k" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.419225 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9adb7fcf-2d97-4c07-85cf-86cef286dc69-catalog-content\") pod \"redhat-operators-t9v8k\" (UID: \"9adb7fcf-2d97-4c07-85cf-86cef286dc69\") " pod="openshift-marketplace/redhat-operators-t9v8k" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.419279 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr6bn\" (UniqueName: \"kubernetes.io/projected/9adb7fcf-2d97-4c07-85cf-86cef286dc69-kube-api-access-qr6bn\") pod \"redhat-operators-t9v8k\" (UID: \"9adb7fcf-2d97-4c07-85cf-86cef286dc69\") " pod="openshift-marketplace/redhat-operators-t9v8k" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.512784 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ca80dd4-6294-4263-b938-188c17814556" path="/var/lib/kubelet/pods/1ca80dd4-6294-4263-b938-188c17814556/volumes" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.514560 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23545a99-b853-4a78-85a6-6186dc3bf903" path="/var/lib/kubelet/pods/23545a99-b853-4a78-85a6-6186dc3bf903/volumes" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.515689 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75521184-74d3-44b9-a410-b0b0fafa4384" path="/var/lib/kubelet/pods/75521184-74d3-44b9-a410-b0b0fafa4384/volumes" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.517247 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8ae1ce1-bf5f-4d13-87b4-0810f63434af" path="/var/lib/kubelet/pods/b8ae1ce1-bf5f-4d13-87b4-0810f63434af/volumes" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.517807 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e74b537e-de07-4a47-85e7-e2bd5a72a11c" path="/var/lib/kubelet/pods/e74b537e-de07-4a47-85e7-e2bd5a72a11c/volumes" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.518516 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pfbfb"] Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.519621 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pfbfb"] Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.519734 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pfbfb" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.520330 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9adb7fcf-2d97-4c07-85cf-86cef286dc69-catalog-content\") pod \"redhat-operators-t9v8k\" (UID: \"9adb7fcf-2d97-4c07-85cf-86cef286dc69\") " pod="openshift-marketplace/redhat-operators-t9v8k" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.520390 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr6bn\" (UniqueName: \"kubernetes.io/projected/9adb7fcf-2d97-4c07-85cf-86cef286dc69-kube-api-access-qr6bn\") pod \"redhat-operators-t9v8k\" (UID: \"9adb7fcf-2d97-4c07-85cf-86cef286dc69\") " pod="openshift-marketplace/redhat-operators-t9v8k" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.520451 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9adb7fcf-2d97-4c07-85cf-86cef286dc69-utilities\") pod \"redhat-operators-t9v8k\" (UID: \"9adb7fcf-2d97-4c07-85cf-86cef286dc69\") " pod="openshift-marketplace/redhat-operators-t9v8k" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.520954 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9adb7fcf-2d97-4c07-85cf-86cef286dc69-utilities\") pod \"redhat-operators-t9v8k\" (UID: \"9adb7fcf-2d97-4c07-85cf-86cef286dc69\") " pod="openshift-marketplace/redhat-operators-t9v8k" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.521217 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9adb7fcf-2d97-4c07-85cf-86cef286dc69-catalog-content\") pod \"redhat-operators-t9v8k\" (UID: \"9adb7fcf-2d97-4c07-85cf-86cef286dc69\") " pod="openshift-marketplace/redhat-operators-t9v8k" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.523272 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.539805 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr6bn\" (UniqueName: \"kubernetes.io/projected/9adb7fcf-2d97-4c07-85cf-86cef286dc69-kube-api-access-qr6bn\") pod \"redhat-operators-t9v8k\" (UID: \"9adb7fcf-2d97-4c07-85cf-86cef286dc69\") " pod="openshift-marketplace/redhat-operators-t9v8k" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.622780 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g9s4\" (UniqueName: \"kubernetes.io/projected/396234a0-0c33-45f1-9e9a-a52c6e69e7c0-kube-api-access-7g9s4\") pod \"redhat-marketplace-pfbfb\" (UID: \"396234a0-0c33-45f1-9e9a-a52c6e69e7c0\") " pod="openshift-marketplace/redhat-marketplace-pfbfb" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.623952 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/396234a0-0c33-45f1-9e9a-a52c6e69e7c0-utilities\") pod \"redhat-marketplace-pfbfb\" (UID: \"396234a0-0c33-45f1-9e9a-a52c6e69e7c0\") " pod="openshift-marketplace/redhat-marketplace-pfbfb" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.624020 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/396234a0-0c33-45f1-9e9a-a52c6e69e7c0-catalog-content\") pod \"redhat-marketplace-pfbfb\" (UID: \"396234a0-0c33-45f1-9e9a-a52c6e69e7c0\") " pod="openshift-marketplace/redhat-marketplace-pfbfb" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.640471 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t9v8k" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.725456 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/396234a0-0c33-45f1-9e9a-a52c6e69e7c0-catalog-content\") pod \"redhat-marketplace-pfbfb\" (UID: \"396234a0-0c33-45f1-9e9a-a52c6e69e7c0\") " pod="openshift-marketplace/redhat-marketplace-pfbfb" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.725552 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g9s4\" (UniqueName: \"kubernetes.io/projected/396234a0-0c33-45f1-9e9a-a52c6e69e7c0-kube-api-access-7g9s4\") pod \"redhat-marketplace-pfbfb\" (UID: \"396234a0-0c33-45f1-9e9a-a52c6e69e7c0\") " pod="openshift-marketplace/redhat-marketplace-pfbfb" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.725595 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/396234a0-0c33-45f1-9e9a-a52c6e69e7c0-utilities\") pod \"redhat-marketplace-pfbfb\" (UID: \"396234a0-0c33-45f1-9e9a-a52c6e69e7c0\") " pod="openshift-marketplace/redhat-marketplace-pfbfb" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.726165 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/396234a0-0c33-45f1-9e9a-a52c6e69e7c0-utilities\") pod \"redhat-marketplace-pfbfb\" (UID: \"396234a0-0c33-45f1-9e9a-a52c6e69e7c0\") " pod="openshift-marketplace/redhat-marketplace-pfbfb" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.726428 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/396234a0-0c33-45f1-9e9a-a52c6e69e7c0-catalog-content\") pod \"redhat-marketplace-pfbfb\" (UID: \"396234a0-0c33-45f1-9e9a-a52c6e69e7c0\") " pod="openshift-marketplace/redhat-marketplace-pfbfb" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.747040 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g9s4\" (UniqueName: \"kubernetes.io/projected/396234a0-0c33-45f1-9e9a-a52c6e69e7c0-kube-api-access-7g9s4\") pod \"redhat-marketplace-pfbfb\" (UID: \"396234a0-0c33-45f1-9e9a-a52c6e69e7c0\") " pod="openshift-marketplace/redhat-marketplace-pfbfb" Oct 13 08:49:09 crc kubenswrapper[4685]: I1013 08:49:09.834516 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pfbfb" Oct 13 08:49:10 crc kubenswrapper[4685]: I1013 08:49:10.089895 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t9v8k"] Oct 13 08:49:10 crc kubenswrapper[4685]: I1013 08:49:10.107758 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pfbfb"] Oct 13 08:49:10 crc kubenswrapper[4685]: I1013 08:49:10.487119 4685 generic.go:334] "Generic (PLEG): container finished" podID="396234a0-0c33-45f1-9e9a-a52c6e69e7c0" containerID="b0531fd83bf9013ac032e996b3754685053bfaef47a509e5293b3941273950a1" exitCode=0 Oct 13 08:49:10 crc kubenswrapper[4685]: I1013 08:49:10.487204 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfbfb" event={"ID":"396234a0-0c33-45f1-9e9a-a52c6e69e7c0","Type":"ContainerDied","Data":"b0531fd83bf9013ac032e996b3754685053bfaef47a509e5293b3941273950a1"} Oct 13 08:49:10 crc kubenswrapper[4685]: I1013 08:49:10.487234 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfbfb" event={"ID":"396234a0-0c33-45f1-9e9a-a52c6e69e7c0","Type":"ContainerStarted","Data":"23f890556d6a8fe0af75ce38018f81c95a770868099bf5c40c3ad1822b62d7cc"} Oct 13 08:49:10 crc kubenswrapper[4685]: I1013 08:49:10.492551 4685 generic.go:334] "Generic (PLEG): container finished" podID="9adb7fcf-2d97-4c07-85cf-86cef286dc69" containerID="be1ce66a0293614fb337db7756ccccb4e859ea8119c7a6b52bf827a369730aea" exitCode=0 Oct 13 08:49:10 crc kubenswrapper[4685]: I1013 08:49:10.492713 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t9v8k" event={"ID":"9adb7fcf-2d97-4c07-85cf-86cef286dc69","Type":"ContainerDied","Data":"be1ce66a0293614fb337db7756ccccb4e859ea8119c7a6b52bf827a369730aea"} Oct 13 08:49:10 crc kubenswrapper[4685]: I1013 08:49:10.492757 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t9v8k" event={"ID":"9adb7fcf-2d97-4c07-85cf-86cef286dc69","Type":"ContainerStarted","Data":"d297479afe0b4fe94428865619e9c07d27c4728ca1dc47e2c1c8b440f4360985"} Oct 13 08:49:11 crc kubenswrapper[4685]: I1013 08:49:11.500822 4685 generic.go:334] "Generic (PLEG): container finished" podID="396234a0-0c33-45f1-9e9a-a52c6e69e7c0" containerID="52e4a1f178e88646c94b919fbe866810a6183d5b45db91f7a6dae5a4e09264ec" exitCode=0 Oct 13 08:49:11 crc kubenswrapper[4685]: I1013 08:49:11.500999 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfbfb" event={"ID":"396234a0-0c33-45f1-9e9a-a52c6e69e7c0","Type":"ContainerDied","Data":"52e4a1f178e88646c94b919fbe866810a6183d5b45db91f7a6dae5a4e09264ec"} Oct 13 08:49:11 crc kubenswrapper[4685]: I1013 08:49:11.708774 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-flhsp"] Oct 13 08:49:11 crc kubenswrapper[4685]: I1013 08:49:11.710033 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-flhsp" Oct 13 08:49:11 crc kubenswrapper[4685]: I1013 08:49:11.712752 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-flhsp"] Oct 13 08:49:11 crc kubenswrapper[4685]: I1013 08:49:11.713100 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 13 08:49:11 crc kubenswrapper[4685]: I1013 08:49:11.792957 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f92f2fa-5809-4485-b5e1-90e6ebf8efd0-catalog-content\") pod \"certified-operators-flhsp\" (UID: \"4f92f2fa-5809-4485-b5e1-90e6ebf8efd0\") " pod="openshift-marketplace/certified-operators-flhsp" Oct 13 08:49:11 crc kubenswrapper[4685]: I1013 08:49:11.793033 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f92f2fa-5809-4485-b5e1-90e6ebf8efd0-utilities\") pod \"certified-operators-flhsp\" (UID: \"4f92f2fa-5809-4485-b5e1-90e6ebf8efd0\") " pod="openshift-marketplace/certified-operators-flhsp" Oct 13 08:49:11 crc kubenswrapper[4685]: I1013 08:49:11.793070 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m4vj\" (UniqueName: \"kubernetes.io/projected/4f92f2fa-5809-4485-b5e1-90e6ebf8efd0-kube-api-access-2m4vj\") pod \"certified-operators-flhsp\" (UID: \"4f92f2fa-5809-4485-b5e1-90e6ebf8efd0\") " pod="openshift-marketplace/certified-operators-flhsp" Oct 13 08:49:11 crc kubenswrapper[4685]: I1013 08:49:11.894072 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m4vj\" (UniqueName: \"kubernetes.io/projected/4f92f2fa-5809-4485-b5e1-90e6ebf8efd0-kube-api-access-2m4vj\") pod \"certified-operators-flhsp\" (UID: \"4f92f2fa-5809-4485-b5e1-90e6ebf8efd0\") " pod="openshift-marketplace/certified-operators-flhsp" Oct 13 08:49:11 crc kubenswrapper[4685]: I1013 08:49:11.894201 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f92f2fa-5809-4485-b5e1-90e6ebf8efd0-catalog-content\") pod \"certified-operators-flhsp\" (UID: \"4f92f2fa-5809-4485-b5e1-90e6ebf8efd0\") " pod="openshift-marketplace/certified-operators-flhsp" Oct 13 08:49:11 crc kubenswrapper[4685]: I1013 08:49:11.894253 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f92f2fa-5809-4485-b5e1-90e6ebf8efd0-utilities\") pod \"certified-operators-flhsp\" (UID: \"4f92f2fa-5809-4485-b5e1-90e6ebf8efd0\") " pod="openshift-marketplace/certified-operators-flhsp" Oct 13 08:49:11 crc kubenswrapper[4685]: I1013 08:49:11.894896 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f92f2fa-5809-4485-b5e1-90e6ebf8efd0-utilities\") pod \"certified-operators-flhsp\" (UID: \"4f92f2fa-5809-4485-b5e1-90e6ebf8efd0\") " pod="openshift-marketplace/certified-operators-flhsp" Oct 13 08:49:11 crc kubenswrapper[4685]: I1013 08:49:11.895025 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f92f2fa-5809-4485-b5e1-90e6ebf8efd0-catalog-content\") pod \"certified-operators-flhsp\" (UID: \"4f92f2fa-5809-4485-b5e1-90e6ebf8efd0\") " pod="openshift-marketplace/certified-operators-flhsp" Oct 13 08:49:11 crc kubenswrapper[4685]: I1013 08:49:11.902705 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9z22p"] Oct 13 08:49:11 crc kubenswrapper[4685]: I1013 08:49:11.904143 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9z22p" Oct 13 08:49:11 crc kubenswrapper[4685]: I1013 08:49:11.907636 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 13 08:49:11 crc kubenswrapper[4685]: I1013 08:49:11.923439 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9z22p"] Oct 13 08:49:11 crc kubenswrapper[4685]: I1013 08:49:11.948648 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m4vj\" (UniqueName: \"kubernetes.io/projected/4f92f2fa-5809-4485-b5e1-90e6ebf8efd0-kube-api-access-2m4vj\") pod \"certified-operators-flhsp\" (UID: \"4f92f2fa-5809-4485-b5e1-90e6ebf8efd0\") " pod="openshift-marketplace/certified-operators-flhsp" Oct 13 08:49:12 crc kubenswrapper[4685]: I1013 08:49:12.038843 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-flhsp" Oct 13 08:49:12 crc kubenswrapper[4685]: I1013 08:49:12.097333 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4dc80bb-b601-4804-80f0-e3f97493b377-utilities\") pod \"community-operators-9z22p\" (UID: \"d4dc80bb-b601-4804-80f0-e3f97493b377\") " pod="openshift-marketplace/community-operators-9z22p" Oct 13 08:49:12 crc kubenswrapper[4685]: I1013 08:49:12.097389 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rd94q\" (UniqueName: \"kubernetes.io/projected/d4dc80bb-b601-4804-80f0-e3f97493b377-kube-api-access-rd94q\") pod \"community-operators-9z22p\" (UID: \"d4dc80bb-b601-4804-80f0-e3f97493b377\") " pod="openshift-marketplace/community-operators-9z22p" Oct 13 08:49:12 crc kubenswrapper[4685]: I1013 08:49:12.097439 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4dc80bb-b601-4804-80f0-e3f97493b377-catalog-content\") pod \"community-operators-9z22p\" (UID: \"d4dc80bb-b601-4804-80f0-e3f97493b377\") " pod="openshift-marketplace/community-operators-9z22p" Oct 13 08:49:12 crc kubenswrapper[4685]: I1013 08:49:12.199319 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4dc80bb-b601-4804-80f0-e3f97493b377-utilities\") pod \"community-operators-9z22p\" (UID: \"d4dc80bb-b601-4804-80f0-e3f97493b377\") " pod="openshift-marketplace/community-operators-9z22p" Oct 13 08:49:12 crc kubenswrapper[4685]: I1013 08:49:12.199769 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rd94q\" (UniqueName: \"kubernetes.io/projected/d4dc80bb-b601-4804-80f0-e3f97493b377-kube-api-access-rd94q\") pod \"community-operators-9z22p\" (UID: \"d4dc80bb-b601-4804-80f0-e3f97493b377\") " pod="openshift-marketplace/community-operators-9z22p" Oct 13 08:49:12 crc kubenswrapper[4685]: I1013 08:49:12.199828 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4dc80bb-b601-4804-80f0-e3f97493b377-catalog-content\") pod \"community-operators-9z22p\" (UID: \"d4dc80bb-b601-4804-80f0-e3f97493b377\") " pod="openshift-marketplace/community-operators-9z22p" Oct 13 08:49:12 crc kubenswrapper[4685]: I1013 08:49:12.199905 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4dc80bb-b601-4804-80f0-e3f97493b377-utilities\") pod \"community-operators-9z22p\" (UID: \"d4dc80bb-b601-4804-80f0-e3f97493b377\") " pod="openshift-marketplace/community-operators-9z22p" Oct 13 08:49:12 crc kubenswrapper[4685]: I1013 08:49:12.200317 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4dc80bb-b601-4804-80f0-e3f97493b377-catalog-content\") pod \"community-operators-9z22p\" (UID: \"d4dc80bb-b601-4804-80f0-e3f97493b377\") " pod="openshift-marketplace/community-operators-9z22p" Oct 13 08:49:12 crc kubenswrapper[4685]: I1013 08:49:12.232725 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rd94q\" (UniqueName: \"kubernetes.io/projected/d4dc80bb-b601-4804-80f0-e3f97493b377-kube-api-access-rd94q\") pod \"community-operators-9z22p\" (UID: \"d4dc80bb-b601-4804-80f0-e3f97493b377\") " pod="openshift-marketplace/community-operators-9z22p" Oct 13 08:49:12 crc kubenswrapper[4685]: I1013 08:49:12.360545 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-flhsp"] Oct 13 08:49:12 crc kubenswrapper[4685]: W1013 08:49:12.367202 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4f92f2fa_5809_4485_b5e1_90e6ebf8efd0.slice/crio-58504542f555a292c16b4db0fb7a71c1c2cf92bfea5b35272040e8329e424560 WatchSource:0}: Error finding container 58504542f555a292c16b4db0fb7a71c1c2cf92bfea5b35272040e8329e424560: Status 404 returned error can't find the container with id 58504542f555a292c16b4db0fb7a71c1c2cf92bfea5b35272040e8329e424560 Oct 13 08:49:12 crc kubenswrapper[4685]: I1013 08:49:12.510299 4685 generic.go:334] "Generic (PLEG): container finished" podID="9adb7fcf-2d97-4c07-85cf-86cef286dc69" containerID="981b1dab9ee49b93ecccd13b39289b3574a531c415a10463190d384dbd3a3ddc" exitCode=0 Oct 13 08:49:12 crc kubenswrapper[4685]: I1013 08:49:12.510729 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t9v8k" event={"ID":"9adb7fcf-2d97-4c07-85cf-86cef286dc69","Type":"ContainerDied","Data":"981b1dab9ee49b93ecccd13b39289b3574a531c415a10463190d384dbd3a3ddc"} Oct 13 08:49:12 crc kubenswrapper[4685]: I1013 08:49:12.516939 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfbfb" event={"ID":"396234a0-0c33-45f1-9e9a-a52c6e69e7c0","Type":"ContainerStarted","Data":"5061dc0da1f0070398705b6a2709012ca7cfb77ada50bcd0af5a5dd686ae5ebe"} Oct 13 08:49:12 crc kubenswrapper[4685]: I1013 08:49:12.522160 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9z22p" Oct 13 08:49:12 crc kubenswrapper[4685]: I1013 08:49:12.522517 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-flhsp" event={"ID":"4f92f2fa-5809-4485-b5e1-90e6ebf8efd0","Type":"ContainerStarted","Data":"e859664bac4b44f9ae676b580cab34dbd01677d6327dbe3469de0ba4f010c3c6"} Oct 13 08:49:12 crc kubenswrapper[4685]: I1013 08:49:12.522549 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-flhsp" event={"ID":"4f92f2fa-5809-4485-b5e1-90e6ebf8efd0","Type":"ContainerStarted","Data":"58504542f555a292c16b4db0fb7a71c1c2cf92bfea5b35272040e8329e424560"} Oct 13 08:49:12 crc kubenswrapper[4685]: I1013 08:49:12.570827 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pfbfb" podStartSLOduration=1.892606845 podStartE2EDuration="3.570803461s" podCreationTimestamp="2025-10-13 08:49:09 +0000 UTC" firstStartedPulling="2025-10-13 08:49:10.490087626 +0000 UTC m=+275.637963387" lastFinishedPulling="2025-10-13 08:49:12.168284202 +0000 UTC m=+277.316160003" observedRunningTime="2025-10-13 08:49:12.55117116 +0000 UTC m=+277.699046931" watchObservedRunningTime="2025-10-13 08:49:12.570803461 +0000 UTC m=+277.718679232" Oct 13 08:49:12 crc kubenswrapper[4685]: I1013 08:49:12.972415 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9z22p"] Oct 13 08:49:13 crc kubenswrapper[4685]: I1013 08:49:13.530029 4685 generic.go:334] "Generic (PLEG): container finished" podID="d4dc80bb-b601-4804-80f0-e3f97493b377" containerID="7c405a80fad14de2e12fc0bd3714b47039f192770677522292a88a61d2ea09d5" exitCode=0 Oct 13 08:49:13 crc kubenswrapper[4685]: I1013 08:49:13.530144 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9z22p" event={"ID":"d4dc80bb-b601-4804-80f0-e3f97493b377","Type":"ContainerDied","Data":"7c405a80fad14de2e12fc0bd3714b47039f192770677522292a88a61d2ea09d5"} Oct 13 08:49:13 crc kubenswrapper[4685]: I1013 08:49:13.530574 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9z22p" event={"ID":"d4dc80bb-b601-4804-80f0-e3f97493b377","Type":"ContainerStarted","Data":"cc386ce3e7fd974632ac31f82b5c6229b5390c5336f79671eea75ad33259bf7e"} Oct 13 08:49:13 crc kubenswrapper[4685]: I1013 08:49:13.538202 4685 generic.go:334] "Generic (PLEG): container finished" podID="4f92f2fa-5809-4485-b5e1-90e6ebf8efd0" containerID="e859664bac4b44f9ae676b580cab34dbd01677d6327dbe3469de0ba4f010c3c6" exitCode=0 Oct 13 08:49:13 crc kubenswrapper[4685]: I1013 08:49:13.538280 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-flhsp" event={"ID":"4f92f2fa-5809-4485-b5e1-90e6ebf8efd0","Type":"ContainerDied","Data":"e859664bac4b44f9ae676b580cab34dbd01677d6327dbe3469de0ba4f010c3c6"} Oct 13 08:49:13 crc kubenswrapper[4685]: I1013 08:49:13.538308 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-flhsp" event={"ID":"4f92f2fa-5809-4485-b5e1-90e6ebf8efd0","Type":"ContainerStarted","Data":"cf36c7324b4a84052655f14b205193ae94c2c8580cabf9bf9561dcde93186a6e"} Oct 13 08:49:13 crc kubenswrapper[4685]: I1013 08:49:13.544135 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t9v8k" event={"ID":"9adb7fcf-2d97-4c07-85cf-86cef286dc69","Type":"ContainerStarted","Data":"1963ef77c2ded741e9369c41c4662a44b06dd234c052104aa801cd8b606449f1"} Oct 13 08:49:13 crc kubenswrapper[4685]: I1013 08:49:13.587044 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t9v8k" podStartSLOduration=2.041931035 podStartE2EDuration="4.587021773s" podCreationTimestamp="2025-10-13 08:49:09 +0000 UTC" firstStartedPulling="2025-10-13 08:49:10.494036131 +0000 UTC m=+275.641911892" lastFinishedPulling="2025-10-13 08:49:13.039126869 +0000 UTC m=+278.187002630" observedRunningTime="2025-10-13 08:49:13.586077257 +0000 UTC m=+278.733953028" watchObservedRunningTime="2025-10-13 08:49:13.587021773 +0000 UTC m=+278.734897534" Oct 13 08:49:14 crc kubenswrapper[4685]: I1013 08:49:14.551027 4685 generic.go:334] "Generic (PLEG): container finished" podID="d4dc80bb-b601-4804-80f0-e3f97493b377" containerID="b5bfd9feaabaec3d4708945a178b949d3b91565d1fad6380d689d171922d36d0" exitCode=0 Oct 13 08:49:14 crc kubenswrapper[4685]: I1013 08:49:14.551123 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9z22p" event={"ID":"d4dc80bb-b601-4804-80f0-e3f97493b377","Type":"ContainerDied","Data":"b5bfd9feaabaec3d4708945a178b949d3b91565d1fad6380d689d171922d36d0"} Oct 13 08:49:14 crc kubenswrapper[4685]: I1013 08:49:14.556768 4685 generic.go:334] "Generic (PLEG): container finished" podID="4f92f2fa-5809-4485-b5e1-90e6ebf8efd0" containerID="cf36c7324b4a84052655f14b205193ae94c2c8580cabf9bf9561dcde93186a6e" exitCode=0 Oct 13 08:49:14 crc kubenswrapper[4685]: I1013 08:49:14.556838 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-flhsp" event={"ID":"4f92f2fa-5809-4485-b5e1-90e6ebf8efd0","Type":"ContainerDied","Data":"cf36c7324b4a84052655f14b205193ae94c2c8580cabf9bf9561dcde93186a6e"} Oct 13 08:49:15 crc kubenswrapper[4685]: I1013 08:49:15.563882 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9z22p" event={"ID":"d4dc80bb-b601-4804-80f0-e3f97493b377","Type":"ContainerStarted","Data":"7241238f3c7cad0aa1e12b333c4237f9ec63623e15a322ce58db21a2da3960ab"} Oct 13 08:49:15 crc kubenswrapper[4685]: I1013 08:49:15.588422 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9z22p" podStartSLOduration=3.1880225859999998 podStartE2EDuration="4.588402328s" podCreationTimestamp="2025-10-13 08:49:11 +0000 UTC" firstStartedPulling="2025-10-13 08:49:13.531923988 +0000 UTC m=+278.679799749" lastFinishedPulling="2025-10-13 08:49:14.93230373 +0000 UTC m=+280.080179491" observedRunningTime="2025-10-13 08:49:15.584274359 +0000 UTC m=+280.732150130" watchObservedRunningTime="2025-10-13 08:49:15.588402328 +0000 UTC m=+280.736278089" Oct 13 08:49:16 crc kubenswrapper[4685]: I1013 08:49:16.571663 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-flhsp" event={"ID":"4f92f2fa-5809-4485-b5e1-90e6ebf8efd0","Type":"ContainerStarted","Data":"7029838667c66663464fa5a5d485f252f32a34618d733e922f63d124ab4f7a37"} Oct 13 08:49:16 crc kubenswrapper[4685]: I1013 08:49:16.589892 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-flhsp" podStartSLOduration=3.016880788 podStartE2EDuration="5.589866878s" podCreationTimestamp="2025-10-13 08:49:11 +0000 UTC" firstStartedPulling="2025-10-13 08:49:12.535095502 +0000 UTC m=+277.682971263" lastFinishedPulling="2025-10-13 08:49:15.108081592 +0000 UTC m=+280.255957353" observedRunningTime="2025-10-13 08:49:16.586620892 +0000 UTC m=+281.734496643" watchObservedRunningTime="2025-10-13 08:49:16.589866878 +0000 UTC m=+281.737742649" Oct 13 08:49:19 crc kubenswrapper[4685]: I1013 08:49:19.641155 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t9v8k" Oct 13 08:49:19 crc kubenswrapper[4685]: I1013 08:49:19.642478 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t9v8k" Oct 13 08:49:19 crc kubenswrapper[4685]: I1013 08:49:19.683502 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t9v8k" Oct 13 08:49:19 crc kubenswrapper[4685]: I1013 08:49:19.835689 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pfbfb" Oct 13 08:49:19 crc kubenswrapper[4685]: I1013 08:49:19.835781 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pfbfb" Oct 13 08:49:19 crc kubenswrapper[4685]: I1013 08:49:19.873044 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pfbfb" Oct 13 08:49:20 crc kubenswrapper[4685]: I1013 08:49:20.644854 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pfbfb" Oct 13 08:49:20 crc kubenswrapper[4685]: I1013 08:49:20.670811 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t9v8k" Oct 13 08:49:22 crc kubenswrapper[4685]: I1013 08:49:22.039803 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-flhsp" Oct 13 08:49:22 crc kubenswrapper[4685]: I1013 08:49:22.040084 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-flhsp" Oct 13 08:49:22 crc kubenswrapper[4685]: I1013 08:49:22.079977 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-flhsp" Oct 13 08:49:22 crc kubenswrapper[4685]: I1013 08:49:22.523096 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9z22p" Oct 13 08:49:22 crc kubenswrapper[4685]: I1013 08:49:22.523175 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9z22p" Oct 13 08:49:22 crc kubenswrapper[4685]: I1013 08:49:22.572138 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9z22p" Oct 13 08:49:22 crc kubenswrapper[4685]: I1013 08:49:22.680740 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9z22p" Oct 13 08:49:22 crc kubenswrapper[4685]: I1013 08:49:22.681903 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-flhsp" Oct 13 08:50:52 crc kubenswrapper[4685]: I1013 08:50:52.980949 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 08:50:52 crc kubenswrapper[4685]: I1013 08:50:52.983088 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 08:51:22 crc kubenswrapper[4685]: I1013 08:51:22.980555 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 08:51:22 crc kubenswrapper[4685]: I1013 08:51:22.981266 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 08:51:52 crc kubenswrapper[4685]: I1013 08:51:52.980138 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 08:51:52 crc kubenswrapper[4685]: I1013 08:51:52.980765 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 08:51:52 crc kubenswrapper[4685]: I1013 08:51:52.980831 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 08:51:52 crc kubenswrapper[4685]: I1013 08:51:52.981730 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ab8bfde2e8129e04a0298b75eb5e60a218ebbe300f76aed267153fb073287fc0"} pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 08:51:52 crc kubenswrapper[4685]: I1013 08:51:52.981823 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" containerID="cri-o://ab8bfde2e8129e04a0298b75eb5e60a218ebbe300f76aed267153fb073287fc0" gracePeriod=600 Oct 13 08:51:53 crc kubenswrapper[4685]: I1013 08:51:53.628464 4685 generic.go:334] "Generic (PLEG): container finished" podID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerID="ab8bfde2e8129e04a0298b75eb5e60a218ebbe300f76aed267153fb073287fc0" exitCode=0 Oct 13 08:51:53 crc kubenswrapper[4685]: I1013 08:51:53.629107 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerDied","Data":"ab8bfde2e8129e04a0298b75eb5e60a218ebbe300f76aed267153fb073287fc0"} Oct 13 08:51:53 crc kubenswrapper[4685]: I1013 08:51:53.629194 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerStarted","Data":"c37641321127c1e407fb4df4bcfeb42705776322c36634a9a7d0dc9c4a308038"} Oct 13 08:51:53 crc kubenswrapper[4685]: I1013 08:51:53.629228 4685 scope.go:117] "RemoveContainer" containerID="d2693d9b84347e956f148ac60adba2f164140ba42fcfa90e75c3bf003a851008" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.715490 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4qcbt"] Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.718809 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.739411 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4qcbt"] Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.765346 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5974fa23-c0f4-4ed7-90ad-9dc7a1646235-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.765433 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7lp6\" (UniqueName: \"kubernetes.io/projected/5974fa23-c0f4-4ed7-90ad-9dc7a1646235-kube-api-access-l7lp6\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.765464 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5974fa23-c0f4-4ed7-90ad-9dc7a1646235-registry-tls\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.765493 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5974fa23-c0f4-4ed7-90ad-9dc7a1646235-registry-certificates\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.765527 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5974fa23-c0f4-4ed7-90ad-9dc7a1646235-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.765591 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5974fa23-c0f4-4ed7-90ad-9dc7a1646235-trusted-ca\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.765627 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5974fa23-c0f4-4ed7-90ad-9dc7a1646235-bound-sa-token\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.765663 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.800838 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.868340 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5974fa23-c0f4-4ed7-90ad-9dc7a1646235-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.868412 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5974fa23-c0f4-4ed7-90ad-9dc7a1646235-trusted-ca\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.868538 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5974fa23-c0f4-4ed7-90ad-9dc7a1646235-bound-sa-token\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.868580 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5974fa23-c0f4-4ed7-90ad-9dc7a1646235-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.868619 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7lp6\" (UniqueName: \"kubernetes.io/projected/5974fa23-c0f4-4ed7-90ad-9dc7a1646235-kube-api-access-l7lp6\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.868646 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5974fa23-c0f4-4ed7-90ad-9dc7a1646235-registry-tls\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.868684 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5974fa23-c0f4-4ed7-90ad-9dc7a1646235-registry-certificates\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.870095 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5974fa23-c0f4-4ed7-90ad-9dc7a1646235-registry-certificates\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.871048 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5974fa23-c0f4-4ed7-90ad-9dc7a1646235-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.872261 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5974fa23-c0f4-4ed7-90ad-9dc7a1646235-trusted-ca\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.880106 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5974fa23-c0f4-4ed7-90ad-9dc7a1646235-registry-tls\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.888235 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5974fa23-c0f4-4ed7-90ad-9dc7a1646235-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.901034 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5974fa23-c0f4-4ed7-90ad-9dc7a1646235-bound-sa-token\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:19 crc kubenswrapper[4685]: I1013 08:52:19.901845 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7lp6\" (UniqueName: \"kubernetes.io/projected/5974fa23-c0f4-4ed7-90ad-9dc7a1646235-kube-api-access-l7lp6\") pod \"image-registry-66df7c8f76-4qcbt\" (UID: \"5974fa23-c0f4-4ed7-90ad-9dc7a1646235\") " pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:20 crc kubenswrapper[4685]: I1013 08:52:20.036314 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:20 crc kubenswrapper[4685]: I1013 08:52:20.279265 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4qcbt"] Oct 13 08:52:20 crc kubenswrapper[4685]: I1013 08:52:20.833502 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" event={"ID":"5974fa23-c0f4-4ed7-90ad-9dc7a1646235","Type":"ContainerStarted","Data":"225f1a714cc8cbc6c5661d4c713fb7839dc2489060633140f45f645ac0d67745"} Oct 13 08:52:20 crc kubenswrapper[4685]: I1013 08:52:20.833552 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" event={"ID":"5974fa23-c0f4-4ed7-90ad-9dc7a1646235","Type":"ContainerStarted","Data":"3bdba928c84d7cd73003d37be69d344c1f351b5c4cf4aca7c5b14bd6e586ce79"} Oct 13 08:52:20 crc kubenswrapper[4685]: I1013 08:52:20.833834 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:20 crc kubenswrapper[4685]: I1013 08:52:20.857034 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" podStartSLOduration=1.857015966 podStartE2EDuration="1.857015966s" podCreationTimestamp="2025-10-13 08:52:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:52:20.854901445 +0000 UTC m=+466.002777246" watchObservedRunningTime="2025-10-13 08:52:20.857015966 +0000 UTC m=+466.004891727" Oct 13 08:52:40 crc kubenswrapper[4685]: I1013 08:52:40.044829 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-4qcbt" Oct 13 08:52:40 crc kubenswrapper[4685]: I1013 08:52:40.109388 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8jzsj"] Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.175850 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" podUID="70fe2241-a755-4556-a84a-733306f90d66" containerName="registry" containerID="cri-o://b5b59afa9d3bc7339ac45f529079822e5e870a3bf5241195fdb3709b89683243" gracePeriod=30 Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.594623 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.735950 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/70fe2241-a755-4556-a84a-733306f90d66-bound-sa-token\") pod \"70fe2241-a755-4556-a84a-733306f90d66\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.737125 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/70fe2241-a755-4556-a84a-733306f90d66-trusted-ca\") pod \"70fe2241-a755-4556-a84a-733306f90d66\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.737242 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/70fe2241-a755-4556-a84a-733306f90d66-ca-trust-extracted\") pod \"70fe2241-a755-4556-a84a-733306f90d66\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.737288 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/70fe2241-a755-4556-a84a-733306f90d66-installation-pull-secrets\") pod \"70fe2241-a755-4556-a84a-733306f90d66\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.737357 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/70fe2241-a755-4556-a84a-733306f90d66-registry-certificates\") pod \"70fe2241-a755-4556-a84a-733306f90d66\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.737391 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/70fe2241-a755-4556-a84a-733306f90d66-registry-tls\") pod \"70fe2241-a755-4556-a84a-733306f90d66\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.737426 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kz5ht\" (UniqueName: \"kubernetes.io/projected/70fe2241-a755-4556-a84a-733306f90d66-kube-api-access-kz5ht\") pod \"70fe2241-a755-4556-a84a-733306f90d66\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.737586 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"70fe2241-a755-4556-a84a-733306f90d66\" (UID: \"70fe2241-a755-4556-a84a-733306f90d66\") " Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.737818 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70fe2241-a755-4556-a84a-733306f90d66-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "70fe2241-a755-4556-a84a-733306f90d66" (UID: "70fe2241-a755-4556-a84a-733306f90d66"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.737986 4685 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/70fe2241-a755-4556-a84a-733306f90d66-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.738002 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70fe2241-a755-4556-a84a-733306f90d66-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "70fe2241-a755-4556-a84a-733306f90d66" (UID: "70fe2241-a755-4556-a84a-733306f90d66"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.746552 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70fe2241-a755-4556-a84a-733306f90d66-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "70fe2241-a755-4556-a84a-733306f90d66" (UID: "70fe2241-a755-4556-a84a-733306f90d66"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.747010 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70fe2241-a755-4556-a84a-733306f90d66-kube-api-access-kz5ht" (OuterVolumeSpecName: "kube-api-access-kz5ht") pod "70fe2241-a755-4556-a84a-733306f90d66" (UID: "70fe2241-a755-4556-a84a-733306f90d66"). InnerVolumeSpecName "kube-api-access-kz5ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.747357 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70fe2241-a755-4556-a84a-733306f90d66-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "70fe2241-a755-4556-a84a-733306f90d66" (UID: "70fe2241-a755-4556-a84a-733306f90d66"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.748690 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70fe2241-a755-4556-a84a-733306f90d66-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "70fe2241-a755-4556-a84a-733306f90d66" (UID: "70fe2241-a755-4556-a84a-733306f90d66"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.749601 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "70fe2241-a755-4556-a84a-733306f90d66" (UID: "70fe2241-a755-4556-a84a-733306f90d66"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.757332 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70fe2241-a755-4556-a84a-733306f90d66-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "70fe2241-a755-4556-a84a-733306f90d66" (UID: "70fe2241-a755-4556-a84a-733306f90d66"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.838988 4685 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/70fe2241-a755-4556-a84a-733306f90d66-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.839053 4685 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/70fe2241-a755-4556-a84a-733306f90d66-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.839073 4685 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/70fe2241-a755-4556-a84a-733306f90d66-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.839118 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kz5ht\" (UniqueName: \"kubernetes.io/projected/70fe2241-a755-4556-a84a-733306f90d66-kube-api-access-kz5ht\") on node \"crc\" DevicePath \"\"" Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.839133 4685 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/70fe2241-a755-4556-a84a-733306f90d66-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 13 08:53:05 crc kubenswrapper[4685]: I1013 08:53:05.839148 4685 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/70fe2241-a755-4556-a84a-733306f90d66-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 13 08:53:06 crc kubenswrapper[4685]: I1013 08:53:06.127680 4685 generic.go:334] "Generic (PLEG): container finished" podID="70fe2241-a755-4556-a84a-733306f90d66" containerID="b5b59afa9d3bc7339ac45f529079822e5e870a3bf5241195fdb3709b89683243" exitCode=0 Oct 13 08:53:06 crc kubenswrapper[4685]: I1013 08:53:06.127727 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" event={"ID":"70fe2241-a755-4556-a84a-733306f90d66","Type":"ContainerDied","Data":"b5b59afa9d3bc7339ac45f529079822e5e870a3bf5241195fdb3709b89683243"} Oct 13 08:53:06 crc kubenswrapper[4685]: I1013 08:53:06.127762 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" event={"ID":"70fe2241-a755-4556-a84a-733306f90d66","Type":"ContainerDied","Data":"51c99adb4fc23a710abd311a3cb9461ff320c80e96227cc77e7fbd53d893481e"} Oct 13 08:53:06 crc kubenswrapper[4685]: I1013 08:53:06.127782 4685 scope.go:117] "RemoveContainer" containerID="b5b59afa9d3bc7339ac45f529079822e5e870a3bf5241195fdb3709b89683243" Oct 13 08:53:06 crc kubenswrapper[4685]: I1013 08:53:06.127841 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8jzsj" Oct 13 08:53:06 crc kubenswrapper[4685]: I1013 08:53:06.147296 4685 scope.go:117] "RemoveContainer" containerID="b5b59afa9d3bc7339ac45f529079822e5e870a3bf5241195fdb3709b89683243" Oct 13 08:53:06 crc kubenswrapper[4685]: E1013 08:53:06.147651 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5b59afa9d3bc7339ac45f529079822e5e870a3bf5241195fdb3709b89683243\": container with ID starting with b5b59afa9d3bc7339ac45f529079822e5e870a3bf5241195fdb3709b89683243 not found: ID does not exist" containerID="b5b59afa9d3bc7339ac45f529079822e5e870a3bf5241195fdb3709b89683243" Oct 13 08:53:06 crc kubenswrapper[4685]: I1013 08:53:06.147680 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5b59afa9d3bc7339ac45f529079822e5e870a3bf5241195fdb3709b89683243"} err="failed to get container status \"b5b59afa9d3bc7339ac45f529079822e5e870a3bf5241195fdb3709b89683243\": rpc error: code = NotFound desc = could not find container \"b5b59afa9d3bc7339ac45f529079822e5e870a3bf5241195fdb3709b89683243\": container with ID starting with b5b59afa9d3bc7339ac45f529079822e5e870a3bf5241195fdb3709b89683243 not found: ID does not exist" Oct 13 08:53:06 crc kubenswrapper[4685]: I1013 08:53:06.161986 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8jzsj"] Oct 13 08:53:06 crc kubenswrapper[4685]: I1013 08:53:06.165733 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8jzsj"] Oct 13 08:53:07 crc kubenswrapper[4685]: I1013 08:53:07.512819 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70fe2241-a755-4556-a84a-733306f90d66" path="/var/lib/kubelet/pods/70fe2241-a755-4556-a84a-733306f90d66/volumes" Oct 13 08:54:22 crc kubenswrapper[4685]: I1013 08:54:22.980359 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 08:54:22 crc kubenswrapper[4685]: I1013 08:54:22.980887 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.214058 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-pw6jw"] Oct 13 08:54:24 crc kubenswrapper[4685]: E1013 08:54:24.214502 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70fe2241-a755-4556-a84a-733306f90d66" containerName="registry" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.214515 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="70fe2241-a755-4556-a84a-733306f90d66" containerName="registry" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.214601 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="70fe2241-a755-4556-a84a-733306f90d66" containerName="registry" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.214952 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-pw6jw" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.219310 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.219335 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.219349 4685 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-xzg5t" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.245334 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-k77gp"] Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.246084 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-k77gp" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.249962 4685 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-rqsc9" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.253307 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-8l97z"] Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.253957 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-8l97z" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.255387 4685 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-8mgfh" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.278540 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-pw6jw"] Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.281843 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-k77gp"] Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.284450 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-8l97z"] Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.384101 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnvl7\" (UniqueName: \"kubernetes.io/projected/1fec02c3-da30-45db-9004-73214b0d5a33-kube-api-access-hnvl7\") pod \"cert-manager-5b446d88c5-k77gp\" (UID: \"1fec02c3-da30-45db-9004-73214b0d5a33\") " pod="cert-manager/cert-manager-5b446d88c5-k77gp" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.384395 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvnvp\" (UniqueName: \"kubernetes.io/projected/17313b92-8f5e-45c2-986c-d86cea130b7e-kube-api-access-fvnvp\") pod \"cert-manager-cainjector-7f985d654d-pw6jw\" (UID: \"17313b92-8f5e-45c2-986c-d86cea130b7e\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-pw6jw" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.384484 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlcc9\" (UniqueName: \"kubernetes.io/projected/d1799c36-72e8-4563-b4df-0115cd0e1108-kube-api-access-xlcc9\") pod \"cert-manager-webhook-5655c58dd6-8l97z\" (UID: \"d1799c36-72e8-4563-b4df-0115cd0e1108\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-8l97z" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.485891 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvnvp\" (UniqueName: \"kubernetes.io/projected/17313b92-8f5e-45c2-986c-d86cea130b7e-kube-api-access-fvnvp\") pod \"cert-manager-cainjector-7f985d654d-pw6jw\" (UID: \"17313b92-8f5e-45c2-986c-d86cea130b7e\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-pw6jw" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.485970 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlcc9\" (UniqueName: \"kubernetes.io/projected/d1799c36-72e8-4563-b4df-0115cd0e1108-kube-api-access-xlcc9\") pod \"cert-manager-webhook-5655c58dd6-8l97z\" (UID: \"d1799c36-72e8-4563-b4df-0115cd0e1108\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-8l97z" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.486017 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnvl7\" (UniqueName: \"kubernetes.io/projected/1fec02c3-da30-45db-9004-73214b0d5a33-kube-api-access-hnvl7\") pod \"cert-manager-5b446d88c5-k77gp\" (UID: \"1fec02c3-da30-45db-9004-73214b0d5a33\") " pod="cert-manager/cert-manager-5b446d88c5-k77gp" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.507890 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnvl7\" (UniqueName: \"kubernetes.io/projected/1fec02c3-da30-45db-9004-73214b0d5a33-kube-api-access-hnvl7\") pod \"cert-manager-5b446d88c5-k77gp\" (UID: \"1fec02c3-da30-45db-9004-73214b0d5a33\") " pod="cert-manager/cert-manager-5b446d88c5-k77gp" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.509879 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlcc9\" (UniqueName: \"kubernetes.io/projected/d1799c36-72e8-4563-b4df-0115cd0e1108-kube-api-access-xlcc9\") pod \"cert-manager-webhook-5655c58dd6-8l97z\" (UID: \"d1799c36-72e8-4563-b4df-0115cd0e1108\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-8l97z" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.521973 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvnvp\" (UniqueName: \"kubernetes.io/projected/17313b92-8f5e-45c2-986c-d86cea130b7e-kube-api-access-fvnvp\") pod \"cert-manager-cainjector-7f985d654d-pw6jw\" (UID: \"17313b92-8f5e-45c2-986c-d86cea130b7e\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-pw6jw" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.530601 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-pw6jw" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.563183 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-k77gp" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.570400 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-8l97z" Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.839806 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-pw6jw"] Oct 13 08:54:24 crc kubenswrapper[4685]: W1013 08:54:24.847387 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17313b92_8f5e_45c2_986c_d86cea130b7e.slice/crio-6302d01ad7d2c9fbcde308fc94e2f8e1c9fd7094d3d65f704adfc8e40b310474 WatchSource:0}: Error finding container 6302d01ad7d2c9fbcde308fc94e2f8e1c9fd7094d3d65f704adfc8e40b310474: Status 404 returned error can't find the container with id 6302d01ad7d2c9fbcde308fc94e2f8e1c9fd7094d3d65f704adfc8e40b310474 Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.853311 4685 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.899193 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-k77gp"] Oct 13 08:54:24 crc kubenswrapper[4685]: W1013 08:54:24.903315 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1fec02c3_da30_45db_9004_73214b0d5a33.slice/crio-50b736c70fb30141303ef026bc5fe1969e5ecd1d0c737acdc50bcfca2819ea5c WatchSource:0}: Error finding container 50b736c70fb30141303ef026bc5fe1969e5ecd1d0c737acdc50bcfca2819ea5c: Status 404 returned error can't find the container with id 50b736c70fb30141303ef026bc5fe1969e5ecd1d0c737acdc50bcfca2819ea5c Oct 13 08:54:24 crc kubenswrapper[4685]: I1013 08:54:24.929493 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-8l97z"] Oct 13 08:54:25 crc kubenswrapper[4685]: I1013 08:54:25.693215 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-k77gp" event={"ID":"1fec02c3-da30-45db-9004-73214b0d5a33","Type":"ContainerStarted","Data":"50b736c70fb30141303ef026bc5fe1969e5ecd1d0c737acdc50bcfca2819ea5c"} Oct 13 08:54:25 crc kubenswrapper[4685]: I1013 08:54:25.695240 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-pw6jw" event={"ID":"17313b92-8f5e-45c2-986c-d86cea130b7e","Type":"ContainerStarted","Data":"6302d01ad7d2c9fbcde308fc94e2f8e1c9fd7094d3d65f704adfc8e40b310474"} Oct 13 08:54:25 crc kubenswrapper[4685]: I1013 08:54:25.696558 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-8l97z" event={"ID":"d1799c36-72e8-4563-b4df-0115cd0e1108","Type":"ContainerStarted","Data":"1abf7c93446e541fbbc1694dfb0219f68dada9f740b9d962f00dbdaa7c8177ba"} Oct 13 08:54:28 crc kubenswrapper[4685]: I1013 08:54:28.714478 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-k77gp" event={"ID":"1fec02c3-da30-45db-9004-73214b0d5a33","Type":"ContainerStarted","Data":"29466a7571bf7bb560f0c23a9767523479445f89e32103e687c147d807bc67ef"} Oct 13 08:54:28 crc kubenswrapper[4685]: I1013 08:54:28.716206 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-pw6jw" event={"ID":"17313b92-8f5e-45c2-986c-d86cea130b7e","Type":"ContainerStarted","Data":"a843b7d4743402ccc91af6e83f5333fc529a55b2f8293a30e3849b02b8465db3"} Oct 13 08:54:28 crc kubenswrapper[4685]: I1013 08:54:28.717595 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-8l97z" event={"ID":"d1799c36-72e8-4563-b4df-0115cd0e1108","Type":"ContainerStarted","Data":"a5fb357031792884db118d68581356038ede7784193acb9ed56c9691148edd9e"} Oct 13 08:54:28 crc kubenswrapper[4685]: I1013 08:54:28.717824 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-8l97z" Oct 13 08:54:28 crc kubenswrapper[4685]: I1013 08:54:28.731112 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-k77gp" podStartSLOduration=1.275473808 podStartE2EDuration="4.731094958s" podCreationTimestamp="2025-10-13 08:54:24 +0000 UTC" firstStartedPulling="2025-10-13 08:54:24.905119075 +0000 UTC m=+590.052994836" lastFinishedPulling="2025-10-13 08:54:28.360740225 +0000 UTC m=+593.508615986" observedRunningTime="2025-10-13 08:54:28.729333689 +0000 UTC m=+593.877209460" watchObservedRunningTime="2025-10-13 08:54:28.731094958 +0000 UTC m=+593.878970719" Oct 13 08:54:28 crc kubenswrapper[4685]: I1013 08:54:28.758667 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-8l97z" podStartSLOduration=1.265643502 podStartE2EDuration="4.758644105s" podCreationTimestamp="2025-10-13 08:54:24 +0000 UTC" firstStartedPulling="2025-10-13 08:54:24.933141305 +0000 UTC m=+590.081017066" lastFinishedPulling="2025-10-13 08:54:28.426141898 +0000 UTC m=+593.574017669" observedRunningTime="2025-10-13 08:54:28.755111805 +0000 UTC m=+593.902987576" watchObservedRunningTime="2025-10-13 08:54:28.758644105 +0000 UTC m=+593.906519886" Oct 13 08:54:28 crc kubenswrapper[4685]: I1013 08:54:28.781706 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-pw6jw" podStartSLOduration=1.2735456250000001 podStartE2EDuration="4.781687765s" podCreationTimestamp="2025-10-13 08:54:24 +0000 UTC" firstStartedPulling="2025-10-13 08:54:24.853102649 +0000 UTC m=+590.000978410" lastFinishedPulling="2025-10-13 08:54:28.361244779 +0000 UTC m=+593.509120550" observedRunningTime="2025-10-13 08:54:28.778859525 +0000 UTC m=+593.926735286" watchObservedRunningTime="2025-10-13 08:54:28.781687765 +0000 UTC m=+593.929563526" Oct 13 08:54:34 crc kubenswrapper[4685]: I1013 08:54:34.574357 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-8l97z" Oct 13 08:54:34 crc kubenswrapper[4685]: I1013 08:54:34.738061 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5kgxk"] Oct 13 08:54:34 crc kubenswrapper[4685]: I1013 08:54:34.739068 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovn-controller" containerID="cri-o://5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226" gracePeriod=30 Oct 13 08:54:34 crc kubenswrapper[4685]: I1013 08:54:34.739180 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="nbdb" containerID="cri-o://da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2" gracePeriod=30 Oct 13 08:54:34 crc kubenswrapper[4685]: I1013 08:54:34.739219 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovn-acl-logging" containerID="cri-o://5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023" gracePeriod=30 Oct 13 08:54:34 crc kubenswrapper[4685]: I1013 08:54:34.739285 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="kube-rbac-proxy-node" containerID="cri-o://84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840" gracePeriod=30 Oct 13 08:54:34 crc kubenswrapper[4685]: I1013 08:54:34.739248 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="northd" containerID="cri-o://c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b" gracePeriod=30 Oct 13 08:54:34 crc kubenswrapper[4685]: I1013 08:54:34.739366 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="sbdb" containerID="cri-o://884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73" gracePeriod=30 Oct 13 08:54:34 crc kubenswrapper[4685]: I1013 08:54:34.739186 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60" gracePeriod=30 Oct 13 08:54:34 crc kubenswrapper[4685]: I1013 08:54:34.804104 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovnkube-controller" containerID="cri-o://379685c6101281fdb70274ed20c9c85e6480bd960816f94d3886064f37a23f5d" gracePeriod=30 Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.086604 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5kgxk_ffa4857b-f741-47d8-b692-74a84be30411/ovnkube-controller/3.log" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.100993 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5kgxk_ffa4857b-f741-47d8-b692-74a84be30411/ovn-acl-logging/0.log" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.101659 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5kgxk_ffa4857b-f741-47d8-b692-74a84be30411/ovn-controller/0.log" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.102148 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.148031 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7cq6j"] Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.148449 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovnkube-controller" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.148511 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovnkube-controller" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.148572 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="kubecfg-setup" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.148624 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="kubecfg-setup" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.148675 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovn-controller" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.148735 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovn-controller" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.148808 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovnkube-controller" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.148874 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovnkube-controller" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.148963 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovn-acl-logging" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.149062 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovn-acl-logging" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.149131 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="northd" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.149192 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="northd" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.149256 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="kube-rbac-proxy-node" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.149318 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="kube-rbac-proxy-node" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.149381 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovnkube-controller" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.149444 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovnkube-controller" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.149508 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovnkube-controller" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.149575 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovnkube-controller" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.149637 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="kube-rbac-proxy-ovn-metrics" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.149700 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="kube-rbac-proxy-ovn-metrics" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.149770 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="sbdb" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.149839 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="sbdb" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.149900 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="nbdb" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.149979 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="nbdb" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.150118 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovn-acl-logging" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.150172 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovnkube-controller" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.150222 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovnkube-controller" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.150272 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="kube-rbac-proxy-ovn-metrics" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.150322 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="nbdb" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.150373 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovn-controller" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.150430 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="sbdb" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.150500 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="northd" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.150564 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="kube-rbac-proxy-node" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.150619 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovnkube-controller" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.150760 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovnkube-controller" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.150815 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovnkube-controller" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.150972 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovnkube-controller" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.151037 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffa4857b-f741-47d8-b692-74a84be30411" containerName="ovnkube-controller" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.152902 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.232989 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ffa4857b-f741-47d8-b692-74a84be30411-ovn-node-metrics-cert\") pod \"ffa4857b-f741-47d8-b692-74a84be30411\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233040 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-cni-bin\") pod \"ffa4857b-f741-47d8-b692-74a84be30411\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233058 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-run-ovn\") pod \"ffa4857b-f741-47d8-b692-74a84be30411\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233080 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-run-ovn-kubernetes\") pod \"ffa4857b-f741-47d8-b692-74a84be30411\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233117 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ffa4857b-f741-47d8-b692-74a84be30411\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233134 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-run-openvswitch\") pod \"ffa4857b-f741-47d8-b692-74a84be30411\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233156 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-etc-openvswitch\") pod \"ffa4857b-f741-47d8-b692-74a84be30411\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233178 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-cni-netd\") pod \"ffa4857b-f741-47d8-b692-74a84be30411\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233196 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-run-systemd\") pod \"ffa4857b-f741-47d8-b692-74a84be30411\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233184 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "ffa4857b-f741-47d8-b692-74a84be30411" (UID: "ffa4857b-f741-47d8-b692-74a84be30411"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233199 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "ffa4857b-f741-47d8-b692-74a84be30411" (UID: "ffa4857b-f741-47d8-b692-74a84be30411"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233256 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "ffa4857b-f741-47d8-b692-74a84be30411" (UID: "ffa4857b-f741-47d8-b692-74a84be30411"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233262 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "ffa4857b-f741-47d8-b692-74a84be30411" (UID: "ffa4857b-f741-47d8-b692-74a84be30411"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233214 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-systemd-units\") pod \"ffa4857b-f741-47d8-b692-74a84be30411\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233282 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "ffa4857b-f741-47d8-b692-74a84be30411" (UID: "ffa4857b-f741-47d8-b692-74a84be30411"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233306 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "ffa4857b-f741-47d8-b692-74a84be30411" (UID: "ffa4857b-f741-47d8-b692-74a84be30411"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233334 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "ffa4857b-f741-47d8-b692-74a84be30411" (UID: "ffa4857b-f741-47d8-b692-74a84be30411"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233350 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-slash\") pod \"ffa4857b-f741-47d8-b692-74a84be30411\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233350 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "ffa4857b-f741-47d8-b692-74a84be30411" (UID: "ffa4857b-f741-47d8-b692-74a84be30411"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233385 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-slash" (OuterVolumeSpecName: "host-slash") pod "ffa4857b-f741-47d8-b692-74a84be30411" (UID: "ffa4857b-f741-47d8-b692-74a84be30411"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233417 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ffa4857b-f741-47d8-b692-74a84be30411-env-overrides\") pod \"ffa4857b-f741-47d8-b692-74a84be30411\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233469 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-log-socket\") pod \"ffa4857b-f741-47d8-b692-74a84be30411\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233510 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-log-socket" (OuterVolumeSpecName: "log-socket") pod "ffa4857b-f741-47d8-b692-74a84be30411" (UID: "ffa4857b-f741-47d8-b692-74a84be30411"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233521 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-kubelet\") pod \"ffa4857b-f741-47d8-b692-74a84be30411\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233553 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "ffa4857b-f741-47d8-b692-74a84be30411" (UID: "ffa4857b-f741-47d8-b692-74a84be30411"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233572 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ffa4857b-f741-47d8-b692-74a84be30411-ovnkube-config\") pod \"ffa4857b-f741-47d8-b692-74a84be30411\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233639 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4hsd\" (UniqueName: \"kubernetes.io/projected/ffa4857b-f741-47d8-b692-74a84be30411-kube-api-access-k4hsd\") pod \"ffa4857b-f741-47d8-b692-74a84be30411\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233687 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-node-log\") pod \"ffa4857b-f741-47d8-b692-74a84be30411\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233742 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-run-netns\") pod \"ffa4857b-f741-47d8-b692-74a84be30411\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233788 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-var-lib-openvswitch\") pod \"ffa4857b-f741-47d8-b692-74a84be30411\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233798 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-node-log" (OuterVolumeSpecName: "node-log") pod "ffa4857b-f741-47d8-b692-74a84be30411" (UID: "ffa4857b-f741-47d8-b692-74a84be30411"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233843 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ffa4857b-f741-47d8-b692-74a84be30411-ovnkube-script-lib\") pod \"ffa4857b-f741-47d8-b692-74a84be30411\" (UID: \"ffa4857b-f741-47d8-b692-74a84be30411\") " Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233861 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "ffa4857b-f741-47d8-b692-74a84be30411" (UID: "ffa4857b-f741-47d8-b692-74a84be30411"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.233907 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "ffa4857b-f741-47d8-b692-74a84be30411" (UID: "ffa4857b-f741-47d8-b692-74a84be30411"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.234213 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-host-run-ovn-kubernetes\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.234326 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-var-lib-openvswitch\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.234423 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffa4857b-f741-47d8-b692-74a84be30411-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "ffa4857b-f741-47d8-b692-74a84be30411" (UID: "ffa4857b-f741-47d8-b692-74a84be30411"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.234439 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.234430 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffa4857b-f741-47d8-b692-74a84be30411-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "ffa4857b-f741-47d8-b692-74a84be30411" (UID: "ffa4857b-f741-47d8-b692-74a84be30411"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.234444 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffa4857b-f741-47d8-b692-74a84be30411-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "ffa4857b-f741-47d8-b692-74a84be30411" (UID: "ffa4857b-f741-47d8-b692-74a84be30411"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.234537 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-ovn-node-metrics-cert\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.234616 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-run-ovn\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.234685 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-node-log\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.234726 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-run-openvswitch\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.234767 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-log-socket\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.234830 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-etc-openvswitch\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.234885 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-host-kubelet\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.235005 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-systemd-units\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.235096 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-ovnkube-script-lib\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.235193 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-env-overrides\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.235262 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-host-cni-netd\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.235324 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-host-slash\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.235368 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-run-systemd\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.235434 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-host-run-netns\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.235499 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jrzp\" (UniqueName: \"kubernetes.io/projected/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-kube-api-access-7jrzp\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.235562 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-ovnkube-config\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.235662 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-host-cni-bin\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.235781 4685 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.235823 4685 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.235849 4685 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.235875 4685 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-slash\") on node \"crc\" DevicePath \"\"" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.235899 4685 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ffa4857b-f741-47d8-b692-74a84be30411-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.235960 4685 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-log-socket\") on node \"crc\" DevicePath \"\"" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.235985 4685 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.236009 4685 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ffa4857b-f741-47d8-b692-74a84be30411-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.236034 4685 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-node-log\") on node \"crc\" DevicePath \"\"" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.236057 4685 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.236082 4685 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.236107 4685 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ffa4857b-f741-47d8-b692-74a84be30411-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.236133 4685 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.236157 4685 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.236181 4685 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.236209 4685 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.236234 4685 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.238819 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffa4857b-f741-47d8-b692-74a84be30411-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "ffa4857b-f741-47d8-b692-74a84be30411" (UID: "ffa4857b-f741-47d8-b692-74a84be30411"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.238950 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffa4857b-f741-47d8-b692-74a84be30411-kube-api-access-k4hsd" (OuterVolumeSpecName: "kube-api-access-k4hsd") pod "ffa4857b-f741-47d8-b692-74a84be30411" (UID: "ffa4857b-f741-47d8-b692-74a84be30411"). InnerVolumeSpecName "kube-api-access-k4hsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.246102 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "ffa4857b-f741-47d8-b692-74a84be30411" (UID: "ffa4857b-f741-47d8-b692-74a84be30411"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.337656 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-host-slash\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.337734 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-run-systemd\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.337777 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-host-run-netns\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.337808 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jrzp\" (UniqueName: \"kubernetes.io/projected/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-kube-api-access-7jrzp\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.337845 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-ovnkube-config\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.337838 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-host-slash\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.337879 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-host-cni-bin\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.337970 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-host-run-netns\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338043 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-host-run-ovn-kubernetes\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338054 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-host-cni-bin\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338078 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-run-systemd\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.337982 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-host-run-ovn-kubernetes\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338267 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-var-lib-openvswitch\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338353 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338412 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-ovn-node-metrics-cert\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338485 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-run-ovn\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338416 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338544 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-node-log\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338601 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-run-openvswitch\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338611 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-node-log\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338626 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-log-socket\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338659 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-etc-openvswitch\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338651 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-run-ovn\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338702 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-log-socket\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338670 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-run-openvswitch\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338730 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-etc-openvswitch\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338685 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-host-kubelet\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338709 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-host-kubelet\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338774 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-systemd-units\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338828 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-ovnkube-script-lib\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338875 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-env-overrides\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338882 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-systemd-units\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338908 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-host-cni-netd\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.338872 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-ovnkube-config\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.339063 4685 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ffa4857b-f741-47d8-b692-74a84be30411-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.339066 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-host-cni-netd\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.339080 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4hsd\" (UniqueName: \"kubernetes.io/projected/ffa4857b-f741-47d8-b692-74a84be30411-kube-api-access-k4hsd\") on node \"crc\" DevicePath \"\"" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.339132 4685 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ffa4857b-f741-47d8-b692-74a84be30411-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.339206 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-var-lib-openvswitch\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.339770 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-env-overrides\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.340183 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-ovnkube-script-lib\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.345587 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-ovn-node-metrics-cert\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.368751 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jrzp\" (UniqueName: \"kubernetes.io/projected/f7ad0d2c-1f86-46f7-9499-e716f9f562c1-kube-api-access-7jrzp\") pod \"ovnkube-node-7cq6j\" (UID: \"f7ad0d2c-1f86-46f7-9499-e716f9f562c1\") " pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.466628 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.666618 4685 scope.go:117] "RemoveContainer" containerID="c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.689220 4685 scope.go:117] "RemoveContainer" containerID="da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.713790 4685 scope.go:117] "RemoveContainer" containerID="e8877e613e54b392abcce5cb52a14b82c09b70d46953871b3b4be7af5863e89b" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.741443 4685 scope.go:117] "RemoveContainer" containerID="379685c6101281fdb70274ed20c9c85e6480bd960816f94d3886064f37a23f5d" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.762571 4685 scope.go:117] "RemoveContainer" containerID="84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.782308 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5kgxk_ffa4857b-f741-47d8-b692-74a84be30411/ovn-acl-logging/0.log" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.783051 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5kgxk_ffa4857b-f741-47d8-b692-74a84be30411/ovn-controller/0.log" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.783690 4685 generic.go:334] "Generic (PLEG): container finished" podID="ffa4857b-f741-47d8-b692-74a84be30411" containerID="884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73" exitCode=0 Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.783727 4685 generic.go:334] "Generic (PLEG): container finished" podID="ffa4857b-f741-47d8-b692-74a84be30411" containerID="06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60" exitCode=0 Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.783735 4685 generic.go:334] "Generic (PLEG): container finished" podID="ffa4857b-f741-47d8-b692-74a84be30411" containerID="84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840" exitCode=0 Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.783774 4685 generic.go:334] "Generic (PLEG): container finished" podID="ffa4857b-f741-47d8-b692-74a84be30411" containerID="5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023" exitCode=143 Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.783785 4685 generic.go:334] "Generic (PLEG): container finished" podID="ffa4857b-f741-47d8-b692-74a84be30411" containerID="5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226" exitCode=143 Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.783825 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.783814 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerDied","Data":"379685c6101281fdb70274ed20c9c85e6480bd960816f94d3886064f37a23f5d"} Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.784050 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerDied","Data":"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73"} Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.784082 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerDied","Data":"da0c65162091232830c912bc6ed1de75137f77d867070fba4348e6807a0546c2"} Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.784101 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerDied","Data":"c530a41f8d94ef9dfca84c0d8c837792bb2f0c9cfd7083be1c32468aa3274f1b"} Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.784116 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerDied","Data":"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60"} Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.784134 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerDied","Data":"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840"} Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.784153 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerDied","Data":"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023"} Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.784149 4685 scope.go:117] "RemoveContainer" containerID="884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.784166 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerDied","Data":"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226"} Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.784303 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kgxk" event={"ID":"ffa4857b-f741-47d8-b692-74a84be30411","Type":"ContainerDied","Data":"c6987d47cba056a20939192576b5ffc101a90da52ee501b1adaa15f600358ddb"} Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.784337 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023"} Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.784359 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226"} Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.784370 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5"} Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.786749 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hg77k_ad75c066-910c-49dc-8e8f-94fe04de919d/kube-multus/2.log" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.787532 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hg77k_ad75c066-910c-49dc-8e8f-94fe04de919d/kube-multus/1.log" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.787605 4685 generic.go:334] "Generic (PLEG): container finished" podID="ad75c066-910c-49dc-8e8f-94fe04de919d" containerID="566c2c013b9b7e7e48b28168baa279368092660f99ed1276e5c6bb9f48db3032" exitCode=2 Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.787681 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hg77k" event={"ID":"ad75c066-910c-49dc-8e8f-94fe04de919d","Type":"ContainerDied","Data":"566c2c013b9b7e7e48b28168baa279368092660f99ed1276e5c6bb9f48db3032"} Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.787716 4685 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d83eb2c651417915faae46704c55e6fd4132f2dfe079333e33b292e847c15c72"} Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.788208 4685 scope.go:117] "RemoveContainer" containerID="566c2c013b9b7e7e48b28168baa279368092660f99ed1276e5c6bb9f48db3032" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.788370 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-hg77k_openshift-multus(ad75c066-910c-49dc-8e8f-94fe04de919d)\"" pod="openshift-multus/multus-hg77k" podUID="ad75c066-910c-49dc-8e8f-94fe04de919d" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.790692 4685 generic.go:334] "Generic (PLEG): container finished" podID="f7ad0d2c-1f86-46f7-9499-e716f9f562c1" containerID="63adf09b9ac0901efab1eed586db35ee7f5b9f0d338391de514825db41b3eb95" exitCode=0 Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.790746 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" event={"ID":"f7ad0d2c-1f86-46f7-9499-e716f9f562c1","Type":"ContainerDied","Data":"63adf09b9ac0901efab1eed586db35ee7f5b9f0d338391de514825db41b3eb95"} Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.790822 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" event={"ID":"f7ad0d2c-1f86-46f7-9499-e716f9f562c1","Type":"ContainerStarted","Data":"40744e53c8e6a0017e1c912b7f6d68b3d9f59696c5fbebc9e6a79e5c24d841fe"} Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.800623 4685 scope.go:117] "RemoveContainer" containerID="0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.819401 4685 scope.go:117] "RemoveContainer" containerID="06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.826540 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5kgxk"] Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.838317 4685 scope.go:117] "RemoveContainer" containerID="5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.839328 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5kgxk"] Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.866351 4685 scope.go:117] "RemoveContainer" containerID="884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.868810 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\": container with ID starting with 884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73 not found: ID does not exist" containerID="884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.868882 4685 kuberuntime_gc.go:150] "Failed to remove container" err="failed to get container status \"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\": rpc error: code = NotFound desc = could not find container \"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\": container with ID starting with 884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73 not found: ID does not exist" containerID="884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.868951 4685 scope.go:117] "RemoveContainer" containerID="06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.874822 4685 scope.go:117] "RemoveContainer" containerID="84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.874997 4685 log.go:32] "RemoveContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to delete container k8s_kube-rbac-proxy-ovn-metrics_ovnkube-node-5kgxk_openshift-ovn-kubernetes_ffa4857b-f741-47d8-b692-74a84be30411_0 in pod sandbox c6987d47cba056a20939192576b5ffc101a90da52ee501b1adaa15f600358ddb from index: no such id: '06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60'" containerID="06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.875051 4685 kuberuntime_gc.go:150] "Failed to remove container" err="rpc error: code = Unknown desc = failed to delete container k8s_kube-rbac-proxy-ovn-metrics_ovnkube-node-5kgxk_openshift-ovn-kubernetes_ffa4857b-f741-47d8-b692-74a84be30411_0 in pod sandbox c6987d47cba056a20939192576b5ffc101a90da52ee501b1adaa15f600358ddb from index: no such id: '06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60'" containerID="06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.875080 4685 scope.go:117] "RemoveContainer" containerID="5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.875137 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\": container with ID starting with 84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840 not found: ID does not exist" containerID="84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.875162 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840"} err="failed to get container status \"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\": rpc error: code = NotFound desc = could not find container \"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\": container with ID starting with 84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.875185 4685 scope.go:117] "RemoveContainer" containerID="5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.926172 4685 scope.go:117] "RemoveContainer" containerID="5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.926409 4685 log.go:32] "RemoveContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to delete container k8s_ovn-acl-logging_ovnkube-node-5kgxk_openshift-ovn-kubernetes_ffa4857b-f741-47d8-b692-74a84be30411_0 in pod sandbox c6987d47cba056a20939192576b5ffc101a90da52ee501b1adaa15f600358ddb from index: no such id: '5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023'" containerID="5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.926445 4685 kuberuntime_gc.go:150] "Failed to remove container" err="rpc error: code = Unknown desc = failed to delete container k8s_ovn-acl-logging_ovnkube-node-5kgxk_openshift-ovn-kubernetes_ffa4857b-f741-47d8-b692-74a84be30411_0 in pod sandbox c6987d47cba056a20939192576b5ffc101a90da52ee501b1adaa15f600358ddb from index: no such id: '5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023'" containerID="5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.926473 4685 scope.go:117] "RemoveContainer" containerID="d83eb2c651417915faae46704c55e6fd4132f2dfe079333e33b292e847c15c72" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.926519 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\": container with ID starting with 5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226 not found: ID does not exist" containerID="5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.926538 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226"} err="failed to get container status \"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\": rpc error: code = NotFound desc = could not find container \"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\": container with ID starting with 5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.926557 4685 scope.go:117] "RemoveContainer" containerID="0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.926856 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\": container with ID starting with 0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5 not found: ID does not exist" containerID="0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.926878 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5"} err="failed to get container status \"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\": rpc error: code = NotFound desc = could not find container \"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\": container with ID starting with 0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.926891 4685 scope.go:117] "RemoveContainer" containerID="884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.927134 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73"} err="failed to get container status \"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\": rpc error: code = NotFound desc = could not find container \"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\": container with ID starting with 884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.927163 4685 scope.go:117] "RemoveContainer" containerID="06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.927427 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\": container with ID starting with 06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60 not found: ID does not exist" containerID="06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.927476 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60"} err="failed to get container status \"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\": rpc error: code = NotFound desc = could not find container \"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\": container with ID starting with 06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.927506 4685 scope.go:117] "RemoveContainer" containerID="84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.927979 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840"} err="failed to get container status \"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\": rpc error: code = NotFound desc = could not find container \"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\": container with ID starting with 84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.928004 4685 scope.go:117] "RemoveContainer" containerID="5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023" Oct 13 08:54:35 crc kubenswrapper[4685]: E1013 08:54:35.928766 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\": container with ID starting with 5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023 not found: ID does not exist" containerID="5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.928818 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023"} err="failed to get container status \"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\": rpc error: code = NotFound desc = could not find container \"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\": container with ID starting with 5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.928838 4685 scope.go:117] "RemoveContainer" containerID="5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.929176 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226"} err="failed to get container status \"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\": rpc error: code = NotFound desc = could not find container \"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\": container with ID starting with 5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.929204 4685 scope.go:117] "RemoveContainer" containerID="0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.929432 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5"} err="failed to get container status \"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\": rpc error: code = NotFound desc = could not find container \"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\": container with ID starting with 0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.929458 4685 scope.go:117] "RemoveContainer" containerID="884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.929717 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73"} err="failed to get container status \"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\": rpc error: code = NotFound desc = could not find container \"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\": container with ID starting with 884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.929742 4685 scope.go:117] "RemoveContainer" containerID="06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.929959 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60"} err="failed to get container status \"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\": rpc error: code = NotFound desc = could not find container \"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\": container with ID starting with 06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.929983 4685 scope.go:117] "RemoveContainer" containerID="84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.930201 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840"} err="failed to get container status \"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\": rpc error: code = NotFound desc = could not find container \"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\": container with ID starting with 84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.930241 4685 scope.go:117] "RemoveContainer" containerID="5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.930405 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023"} err="failed to get container status \"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\": rpc error: code = NotFound desc = could not find container \"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\": container with ID starting with 5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.930423 4685 scope.go:117] "RemoveContainer" containerID="5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.930608 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226"} err="failed to get container status \"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\": rpc error: code = NotFound desc = could not find container \"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\": container with ID starting with 5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.930649 4685 scope.go:117] "RemoveContainer" containerID="0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.930888 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5"} err="failed to get container status \"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\": rpc error: code = NotFound desc = could not find container \"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\": container with ID starting with 0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.930962 4685 scope.go:117] "RemoveContainer" containerID="884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.931306 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73"} err="failed to get container status \"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\": rpc error: code = NotFound desc = could not find container \"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\": container with ID starting with 884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.931327 4685 scope.go:117] "RemoveContainer" containerID="06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.931510 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60"} err="failed to get container status \"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\": rpc error: code = NotFound desc = could not find container \"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\": container with ID starting with 06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.931536 4685 scope.go:117] "RemoveContainer" containerID="84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.931709 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840"} err="failed to get container status \"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\": rpc error: code = NotFound desc = could not find container \"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\": container with ID starting with 84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.931733 4685 scope.go:117] "RemoveContainer" containerID="5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.931971 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023"} err="failed to get container status \"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\": rpc error: code = NotFound desc = could not find container \"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\": container with ID starting with 5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.931997 4685 scope.go:117] "RemoveContainer" containerID="5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.932183 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226"} err="failed to get container status \"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\": rpc error: code = NotFound desc = could not find container \"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\": container with ID starting with 5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.932393 4685 scope.go:117] "RemoveContainer" containerID="0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.932702 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5"} err="failed to get container status \"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\": rpc error: code = NotFound desc = could not find container \"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\": container with ID starting with 0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.932737 4685 scope.go:117] "RemoveContainer" containerID="884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.932970 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73"} err="failed to get container status \"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\": rpc error: code = NotFound desc = could not find container \"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\": container with ID starting with 884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.932993 4685 scope.go:117] "RemoveContainer" containerID="06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.933187 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60"} err="failed to get container status \"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\": rpc error: code = NotFound desc = could not find container \"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\": container with ID starting with 06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.933208 4685 scope.go:117] "RemoveContainer" containerID="84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.933395 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840"} err="failed to get container status \"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\": rpc error: code = NotFound desc = could not find container \"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\": container with ID starting with 84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.933417 4685 scope.go:117] "RemoveContainer" containerID="5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.933620 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023"} err="failed to get container status \"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\": rpc error: code = NotFound desc = could not find container \"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\": container with ID starting with 5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.933648 4685 scope.go:117] "RemoveContainer" containerID="5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.933839 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226"} err="failed to get container status \"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\": rpc error: code = NotFound desc = could not find container \"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\": container with ID starting with 5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.933863 4685 scope.go:117] "RemoveContainer" containerID="0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.934119 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5"} err="failed to get container status \"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\": rpc error: code = NotFound desc = could not find container \"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\": container with ID starting with 0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.934139 4685 scope.go:117] "RemoveContainer" containerID="884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.934429 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73"} err="failed to get container status \"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\": rpc error: code = NotFound desc = could not find container \"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\": container with ID starting with 884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.934459 4685 scope.go:117] "RemoveContainer" containerID="06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.934661 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60"} err="failed to get container status \"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\": rpc error: code = NotFound desc = could not find container \"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\": container with ID starting with 06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.934687 4685 scope.go:117] "RemoveContainer" containerID="84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.934878 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840"} err="failed to get container status \"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\": rpc error: code = NotFound desc = could not find container \"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\": container with ID starting with 84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.934903 4685 scope.go:117] "RemoveContainer" containerID="5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.935112 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023"} err="failed to get container status \"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\": rpc error: code = NotFound desc = could not find container \"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\": container with ID starting with 5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.935133 4685 scope.go:117] "RemoveContainer" containerID="5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.935338 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226"} err="failed to get container status \"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\": rpc error: code = NotFound desc = could not find container \"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\": container with ID starting with 5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.935362 4685 scope.go:117] "RemoveContainer" containerID="0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.935545 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5"} err="failed to get container status \"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\": rpc error: code = NotFound desc = could not find container \"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\": container with ID starting with 0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.935565 4685 scope.go:117] "RemoveContainer" containerID="884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.935758 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73"} err="failed to get container status \"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\": rpc error: code = NotFound desc = could not find container \"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\": container with ID starting with 884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.935782 4685 scope.go:117] "RemoveContainer" containerID="06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.935982 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60"} err="failed to get container status \"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\": rpc error: code = NotFound desc = could not find container \"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\": container with ID starting with 06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.936032 4685 scope.go:117] "RemoveContainer" containerID="84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.936266 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840"} err="failed to get container status \"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\": rpc error: code = NotFound desc = could not find container \"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\": container with ID starting with 84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.936333 4685 scope.go:117] "RemoveContainer" containerID="5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.936587 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023"} err="failed to get container status \"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\": rpc error: code = NotFound desc = could not find container \"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\": container with ID starting with 5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.936633 4685 scope.go:117] "RemoveContainer" containerID="5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.936856 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226"} err="failed to get container status \"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\": rpc error: code = NotFound desc = could not find container \"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\": container with ID starting with 5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.936883 4685 scope.go:117] "RemoveContainer" containerID="0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.937213 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5"} err="failed to get container status \"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\": rpc error: code = NotFound desc = could not find container \"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\": container with ID starting with 0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.937232 4685 scope.go:117] "RemoveContainer" containerID="884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.937628 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73"} err="failed to get container status \"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\": rpc error: code = NotFound desc = could not find container \"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\": container with ID starting with 884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.937649 4685 scope.go:117] "RemoveContainer" containerID="06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.937843 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60"} err="failed to get container status \"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\": rpc error: code = NotFound desc = could not find container \"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\": container with ID starting with 06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.937868 4685 scope.go:117] "RemoveContainer" containerID="84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.938182 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840"} err="failed to get container status \"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\": rpc error: code = NotFound desc = could not find container \"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\": container with ID starting with 84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.938236 4685 scope.go:117] "RemoveContainer" containerID="5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.938541 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023"} err="failed to get container status \"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\": rpc error: code = NotFound desc = could not find container \"5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023\": container with ID starting with 5bebc3141579f0d1f4f2fb170d1d56e00563fefa3d07d42d1746854fbb6aa023 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.938576 4685 scope.go:117] "RemoveContainer" containerID="5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.938900 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226"} err="failed to get container status \"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\": rpc error: code = NotFound desc = could not find container \"5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226\": container with ID starting with 5606d6459f2566f1c42abc3724c1c4ff28bd8c56b33ccf98fe3795d9ed28d226 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.938938 4685 scope.go:117] "RemoveContainer" containerID="0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.939142 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5"} err="failed to get container status \"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\": rpc error: code = NotFound desc = could not find container \"0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5\": container with ID starting with 0b8254ca04bfbc3d56f9283787372b2bf2cf9294b7d39174c4c7ca7142170ce5 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.939164 4685 scope.go:117] "RemoveContainer" containerID="884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.939449 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73"} err="failed to get container status \"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\": rpc error: code = NotFound desc = could not find container \"884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73\": container with ID starting with 884d63c6fba3fe61f0dfb76008c827d65eda9fab2b5f72c53f6c7494253d7f73 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.939472 4685 scope.go:117] "RemoveContainer" containerID="06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.939771 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60"} err="failed to get container status \"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\": rpc error: code = NotFound desc = could not find container \"06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60\": container with ID starting with 06d3d2b548c8b1d6cf215f3eb34922b49385b98021a3e9fc95177b89a8fbba60 not found: ID does not exist" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.939818 4685 scope.go:117] "RemoveContainer" containerID="84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840" Oct 13 08:54:35 crc kubenswrapper[4685]: I1013 08:54:35.940357 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840"} err="failed to get container status \"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\": rpc error: code = NotFound desc = could not find container \"84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840\": container with ID starting with 84f647bd75677fedfdcc91733c24face6e23abefff835fe849ccfa41fb9e7840 not found: ID does not exist" Oct 13 08:54:36 crc kubenswrapper[4685]: I1013 08:54:36.797938 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hg77k_ad75c066-910c-49dc-8e8f-94fe04de919d/kube-multus/2.log" Oct 13 08:54:36 crc kubenswrapper[4685]: I1013 08:54:36.803165 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" event={"ID":"f7ad0d2c-1f86-46f7-9499-e716f9f562c1","Type":"ContainerStarted","Data":"ed2f8d53dc786316b0d9d6fd4bacd6758c8bff60174d40ee0721d5214822a20f"} Oct 13 08:54:36 crc kubenswrapper[4685]: I1013 08:54:36.803213 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" event={"ID":"f7ad0d2c-1f86-46f7-9499-e716f9f562c1","Type":"ContainerStarted","Data":"e6af219d9c348e0e1d47088a603d332033098f4109ec8b4ed2a3b58b7d76e1d1"} Oct 13 08:54:36 crc kubenswrapper[4685]: I1013 08:54:36.803225 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" event={"ID":"f7ad0d2c-1f86-46f7-9499-e716f9f562c1","Type":"ContainerStarted","Data":"780b30a1edbe12058181b60495677734c51f441469d6ddfe16a13dc569424f8b"} Oct 13 08:54:36 crc kubenswrapper[4685]: I1013 08:54:36.803235 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" event={"ID":"f7ad0d2c-1f86-46f7-9499-e716f9f562c1","Type":"ContainerStarted","Data":"a07ca5c634b697e606da666082471b2bd8cb82787f17810c3df3e639c3dd9a0b"} Oct 13 08:54:36 crc kubenswrapper[4685]: I1013 08:54:36.803244 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" event={"ID":"f7ad0d2c-1f86-46f7-9499-e716f9f562c1","Type":"ContainerStarted","Data":"374aab514bbbdf908d87222fe9ca7f10446752c3946aed306d179d76def2f5bb"} Oct 13 08:54:36 crc kubenswrapper[4685]: I1013 08:54:36.803257 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" event={"ID":"f7ad0d2c-1f86-46f7-9499-e716f9f562c1","Type":"ContainerStarted","Data":"6f61a7d1c08b1078a3e2600fd297109d95d32ec7ebadfaaca6728bd69faa55ea"} Oct 13 08:54:37 crc kubenswrapper[4685]: I1013 08:54:37.515070 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffa4857b-f741-47d8-b692-74a84be30411" path="/var/lib/kubelet/pods/ffa4857b-f741-47d8-b692-74a84be30411/volumes" Oct 13 08:54:38 crc kubenswrapper[4685]: I1013 08:54:38.818983 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" event={"ID":"f7ad0d2c-1f86-46f7-9499-e716f9f562c1","Type":"ContainerStarted","Data":"98bab44e9dcb395a5964f35b0192b66e6e8abc82e58a46a6412aa6f7e98ebf8b"} Oct 13 08:54:41 crc kubenswrapper[4685]: I1013 08:54:41.840717 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" event={"ID":"f7ad0d2c-1f86-46f7-9499-e716f9f562c1","Type":"ContainerStarted","Data":"b3cacb9084899d3b2da55966e671beb4573419c1258cd4cacd361c2b3663cfcc"} Oct 13 08:54:41 crc kubenswrapper[4685]: I1013 08:54:41.841371 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:41 crc kubenswrapper[4685]: I1013 08:54:41.841389 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:41 crc kubenswrapper[4685]: I1013 08:54:41.841399 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:41 crc kubenswrapper[4685]: I1013 08:54:41.874532 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" podStartSLOduration=6.874489682 podStartE2EDuration="6.874489682s" podCreationTimestamp="2025-10-13 08:54:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:54:41.871239481 +0000 UTC m=+607.019115252" watchObservedRunningTime="2025-10-13 08:54:41.874489682 +0000 UTC m=+607.022365453" Oct 13 08:54:41 crc kubenswrapper[4685]: I1013 08:54:41.892161 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:41 crc kubenswrapper[4685]: I1013 08:54:41.892358 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:54:46 crc kubenswrapper[4685]: I1013 08:54:46.502566 4685 scope.go:117] "RemoveContainer" containerID="566c2c013b9b7e7e48b28168baa279368092660f99ed1276e5c6bb9f48db3032" Oct 13 08:54:46 crc kubenswrapper[4685]: E1013 08:54:46.504993 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-hg77k_openshift-multus(ad75c066-910c-49dc-8e8f-94fe04de919d)\"" pod="openshift-multus/multus-hg77k" podUID="ad75c066-910c-49dc-8e8f-94fe04de919d" Oct 13 08:54:52 crc kubenswrapper[4685]: I1013 08:54:52.980672 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 08:54:52 crc kubenswrapper[4685]: I1013 08:54:52.981041 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 08:54:58 crc kubenswrapper[4685]: I1013 08:54:58.502467 4685 scope.go:117] "RemoveContainer" containerID="566c2c013b9b7e7e48b28168baa279368092660f99ed1276e5c6bb9f48db3032" Oct 13 08:54:58 crc kubenswrapper[4685]: I1013 08:54:58.949677 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hg77k_ad75c066-910c-49dc-8e8f-94fe04de919d/kube-multus/2.log" Oct 13 08:54:58 crc kubenswrapper[4685]: I1013 08:54:58.950156 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hg77k" event={"ID":"ad75c066-910c-49dc-8e8f-94fe04de919d","Type":"ContainerStarted","Data":"3b9a7973f5bc45241aa67319fa376278ac2bf5ab27cab5f883a86c5c5250ab7c"} Oct 13 08:55:05 crc kubenswrapper[4685]: I1013 08:55:05.498964 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7cq6j" Oct 13 08:55:15 crc kubenswrapper[4685]: I1013 08:55:15.296562 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc"] Oct 13 08:55:15 crc kubenswrapper[4685]: I1013 08:55:15.298422 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc" Oct 13 08:55:15 crc kubenswrapper[4685]: I1013 08:55:15.302345 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd58m\" (UniqueName: \"kubernetes.io/projected/207d7e9f-5cfb-4476-962f-8f6b960d32d2-kube-api-access-gd58m\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc\" (UID: \"207d7e9f-5cfb-4476-962f-8f6b960d32d2\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc" Oct 13 08:55:15 crc kubenswrapper[4685]: I1013 08:55:15.302511 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/207d7e9f-5cfb-4476-962f-8f6b960d32d2-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc\" (UID: \"207d7e9f-5cfb-4476-962f-8f6b960d32d2\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc" Oct 13 08:55:15 crc kubenswrapper[4685]: I1013 08:55:15.302578 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/207d7e9f-5cfb-4476-962f-8f6b960d32d2-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc\" (UID: \"207d7e9f-5cfb-4476-962f-8f6b960d32d2\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc" Oct 13 08:55:15 crc kubenswrapper[4685]: I1013 08:55:15.302364 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 13 08:55:15 crc kubenswrapper[4685]: I1013 08:55:15.318182 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc"] Oct 13 08:55:15 crc kubenswrapper[4685]: I1013 08:55:15.404166 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/207d7e9f-5cfb-4476-962f-8f6b960d32d2-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc\" (UID: \"207d7e9f-5cfb-4476-962f-8f6b960d32d2\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc" Oct 13 08:55:15 crc kubenswrapper[4685]: I1013 08:55:15.404230 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/207d7e9f-5cfb-4476-962f-8f6b960d32d2-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc\" (UID: \"207d7e9f-5cfb-4476-962f-8f6b960d32d2\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc" Oct 13 08:55:15 crc kubenswrapper[4685]: I1013 08:55:15.404283 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd58m\" (UniqueName: \"kubernetes.io/projected/207d7e9f-5cfb-4476-962f-8f6b960d32d2-kube-api-access-gd58m\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc\" (UID: \"207d7e9f-5cfb-4476-962f-8f6b960d32d2\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc" Oct 13 08:55:15 crc kubenswrapper[4685]: I1013 08:55:15.404736 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/207d7e9f-5cfb-4476-962f-8f6b960d32d2-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc\" (UID: \"207d7e9f-5cfb-4476-962f-8f6b960d32d2\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc" Oct 13 08:55:15 crc kubenswrapper[4685]: I1013 08:55:15.405009 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/207d7e9f-5cfb-4476-962f-8f6b960d32d2-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc\" (UID: \"207d7e9f-5cfb-4476-962f-8f6b960d32d2\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc" Oct 13 08:55:15 crc kubenswrapper[4685]: I1013 08:55:15.427563 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd58m\" (UniqueName: \"kubernetes.io/projected/207d7e9f-5cfb-4476-962f-8f6b960d32d2-kube-api-access-gd58m\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc\" (UID: \"207d7e9f-5cfb-4476-962f-8f6b960d32d2\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc" Oct 13 08:55:15 crc kubenswrapper[4685]: I1013 08:55:15.622341 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc" Oct 13 08:55:16 crc kubenswrapper[4685]: I1013 08:55:16.098183 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc"] Oct 13 08:55:17 crc kubenswrapper[4685]: I1013 08:55:17.062121 4685 generic.go:334] "Generic (PLEG): container finished" podID="207d7e9f-5cfb-4476-962f-8f6b960d32d2" containerID="f373f84eb391812f14a466c09b6bddfd1c6d40a0cabe119f0e770d608c561baa" exitCode=0 Oct 13 08:55:17 crc kubenswrapper[4685]: I1013 08:55:17.062195 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc" event={"ID":"207d7e9f-5cfb-4476-962f-8f6b960d32d2","Type":"ContainerDied","Data":"f373f84eb391812f14a466c09b6bddfd1c6d40a0cabe119f0e770d608c561baa"} Oct 13 08:55:17 crc kubenswrapper[4685]: I1013 08:55:17.062257 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc" event={"ID":"207d7e9f-5cfb-4476-962f-8f6b960d32d2","Type":"ContainerStarted","Data":"ea05d1e4429a83964214bb1f42873afc915c4b4a4277a3cb56b429b36d243d40"} Oct 13 08:55:20 crc kubenswrapper[4685]: I1013 08:55:20.082294 4685 generic.go:334] "Generic (PLEG): container finished" podID="207d7e9f-5cfb-4476-962f-8f6b960d32d2" containerID="8824e03cafcaa14867d9d7db4027ade90999646ecbefe7a1cb9ce026c914e9fe" exitCode=0 Oct 13 08:55:20 crc kubenswrapper[4685]: I1013 08:55:20.082451 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc" event={"ID":"207d7e9f-5cfb-4476-962f-8f6b960d32d2","Type":"ContainerDied","Data":"8824e03cafcaa14867d9d7db4027ade90999646ecbefe7a1cb9ce026c914e9fe"} Oct 13 08:55:21 crc kubenswrapper[4685]: I1013 08:55:21.093159 4685 generic.go:334] "Generic (PLEG): container finished" podID="207d7e9f-5cfb-4476-962f-8f6b960d32d2" containerID="2ef7c9ed719edc877cb23e84d3fa798e94345cd237acb582e1fb744308eaf5e6" exitCode=0 Oct 13 08:55:21 crc kubenswrapper[4685]: I1013 08:55:21.093226 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc" event={"ID":"207d7e9f-5cfb-4476-962f-8f6b960d32d2","Type":"ContainerDied","Data":"2ef7c9ed719edc877cb23e84d3fa798e94345cd237acb582e1fb744308eaf5e6"} Oct 13 08:55:22 crc kubenswrapper[4685]: I1013 08:55:22.428591 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc" Oct 13 08:55:22 crc kubenswrapper[4685]: I1013 08:55:22.515858 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/207d7e9f-5cfb-4476-962f-8f6b960d32d2-bundle\") pod \"207d7e9f-5cfb-4476-962f-8f6b960d32d2\" (UID: \"207d7e9f-5cfb-4476-962f-8f6b960d32d2\") " Oct 13 08:55:22 crc kubenswrapper[4685]: I1013 08:55:22.517017 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/207d7e9f-5cfb-4476-962f-8f6b960d32d2-bundle" (OuterVolumeSpecName: "bundle") pod "207d7e9f-5cfb-4476-962f-8f6b960d32d2" (UID: "207d7e9f-5cfb-4476-962f-8f6b960d32d2"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:55:22 crc kubenswrapper[4685]: I1013 08:55:22.517714 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gd58m\" (UniqueName: \"kubernetes.io/projected/207d7e9f-5cfb-4476-962f-8f6b960d32d2-kube-api-access-gd58m\") pod \"207d7e9f-5cfb-4476-962f-8f6b960d32d2\" (UID: \"207d7e9f-5cfb-4476-962f-8f6b960d32d2\") " Oct 13 08:55:22 crc kubenswrapper[4685]: I1013 08:55:22.518010 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/207d7e9f-5cfb-4476-962f-8f6b960d32d2-util\") pod \"207d7e9f-5cfb-4476-962f-8f6b960d32d2\" (UID: \"207d7e9f-5cfb-4476-962f-8f6b960d32d2\") " Oct 13 08:55:22 crc kubenswrapper[4685]: I1013 08:55:22.518564 4685 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/207d7e9f-5cfb-4476-962f-8f6b960d32d2-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 08:55:22 crc kubenswrapper[4685]: I1013 08:55:22.526836 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/207d7e9f-5cfb-4476-962f-8f6b960d32d2-kube-api-access-gd58m" (OuterVolumeSpecName: "kube-api-access-gd58m") pod "207d7e9f-5cfb-4476-962f-8f6b960d32d2" (UID: "207d7e9f-5cfb-4476-962f-8f6b960d32d2"). InnerVolumeSpecName "kube-api-access-gd58m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:55:22 crc kubenswrapper[4685]: I1013 08:55:22.542117 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/207d7e9f-5cfb-4476-962f-8f6b960d32d2-util" (OuterVolumeSpecName: "util") pod "207d7e9f-5cfb-4476-962f-8f6b960d32d2" (UID: "207d7e9f-5cfb-4476-962f-8f6b960d32d2"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:55:22 crc kubenswrapper[4685]: I1013 08:55:22.620095 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gd58m\" (UniqueName: \"kubernetes.io/projected/207d7e9f-5cfb-4476-962f-8f6b960d32d2-kube-api-access-gd58m\") on node \"crc\" DevicePath \"\"" Oct 13 08:55:22 crc kubenswrapper[4685]: I1013 08:55:22.620138 4685 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/207d7e9f-5cfb-4476-962f-8f6b960d32d2-util\") on node \"crc\" DevicePath \"\"" Oct 13 08:55:22 crc kubenswrapper[4685]: I1013 08:55:22.980010 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 08:55:22 crc kubenswrapper[4685]: I1013 08:55:22.980142 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 08:55:22 crc kubenswrapper[4685]: I1013 08:55:22.980227 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 08:55:22 crc kubenswrapper[4685]: I1013 08:55:22.981422 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c37641321127c1e407fb4df4bcfeb42705776322c36634a9a7d0dc9c4a308038"} pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 08:55:22 crc kubenswrapper[4685]: I1013 08:55:22.981579 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" containerID="cri-o://c37641321127c1e407fb4df4bcfeb42705776322c36634a9a7d0dc9c4a308038" gracePeriod=600 Oct 13 08:55:23 crc kubenswrapper[4685]: I1013 08:55:23.116117 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc" event={"ID":"207d7e9f-5cfb-4476-962f-8f6b960d32d2","Type":"ContainerDied","Data":"ea05d1e4429a83964214bb1f42873afc915c4b4a4277a3cb56b429b36d243d40"} Oct 13 08:55:23 crc kubenswrapper[4685]: I1013 08:55:23.116165 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc" Oct 13 08:55:23 crc kubenswrapper[4685]: I1013 08:55:23.116186 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea05d1e4429a83964214bb1f42873afc915c4b4a4277a3cb56b429b36d243d40" Oct 13 08:55:23 crc kubenswrapper[4685]: I1013 08:55:23.133423 4685 generic.go:334] "Generic (PLEG): container finished" podID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerID="c37641321127c1e407fb4df4bcfeb42705776322c36634a9a7d0dc9c4a308038" exitCode=0 Oct 13 08:55:23 crc kubenswrapper[4685]: I1013 08:55:23.133510 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerDied","Data":"c37641321127c1e407fb4df4bcfeb42705776322c36634a9a7d0dc9c4a308038"} Oct 13 08:55:23 crc kubenswrapper[4685]: I1013 08:55:23.133592 4685 scope.go:117] "RemoveContainer" containerID="ab8bfde2e8129e04a0298b75eb5e60a218ebbe300f76aed267153fb073287fc0" Oct 13 08:55:24 crc kubenswrapper[4685]: I1013 08:55:24.145028 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerStarted","Data":"c54513235e556be91a1895a9b0943234dc7b6e87cbf946c33f9dc14ea2f61818"} Oct 13 08:55:26 crc kubenswrapper[4685]: I1013 08:55:26.861372 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-7l6n7"] Oct 13 08:55:26 crc kubenswrapper[4685]: E1013 08:55:26.862238 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="207d7e9f-5cfb-4476-962f-8f6b960d32d2" containerName="util" Oct 13 08:55:26 crc kubenswrapper[4685]: I1013 08:55:26.862256 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="207d7e9f-5cfb-4476-962f-8f6b960d32d2" containerName="util" Oct 13 08:55:26 crc kubenswrapper[4685]: E1013 08:55:26.862266 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="207d7e9f-5cfb-4476-962f-8f6b960d32d2" containerName="pull" Oct 13 08:55:26 crc kubenswrapper[4685]: I1013 08:55:26.862272 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="207d7e9f-5cfb-4476-962f-8f6b960d32d2" containerName="pull" Oct 13 08:55:26 crc kubenswrapper[4685]: E1013 08:55:26.862287 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="207d7e9f-5cfb-4476-962f-8f6b960d32d2" containerName="extract" Oct 13 08:55:26 crc kubenswrapper[4685]: I1013 08:55:26.862294 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="207d7e9f-5cfb-4476-962f-8f6b960d32d2" containerName="extract" Oct 13 08:55:26 crc kubenswrapper[4685]: I1013 08:55:26.862399 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="207d7e9f-5cfb-4476-962f-8f6b960d32d2" containerName="extract" Oct 13 08:55:26 crc kubenswrapper[4685]: I1013 08:55:26.862858 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-7l6n7" Oct 13 08:55:26 crc kubenswrapper[4685]: I1013 08:55:26.864555 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 13 08:55:26 crc kubenswrapper[4685]: I1013 08:55:26.866498 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 13 08:55:26 crc kubenswrapper[4685]: I1013 08:55:26.867349 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-jbbgp" Oct 13 08:55:26 crc kubenswrapper[4685]: I1013 08:55:26.878868 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfbj7\" (UniqueName: \"kubernetes.io/projected/4d407a70-cfdb-427d-94a8-c975db126733-kube-api-access-cfbj7\") pod \"nmstate-operator-858ddd8f98-7l6n7\" (UID: \"4d407a70-cfdb-427d-94a8-c975db126733\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-7l6n7" Oct 13 08:55:26 crc kubenswrapper[4685]: I1013 08:55:26.881226 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-7l6n7"] Oct 13 08:55:26 crc kubenswrapper[4685]: I1013 08:55:26.980196 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfbj7\" (UniqueName: \"kubernetes.io/projected/4d407a70-cfdb-427d-94a8-c975db126733-kube-api-access-cfbj7\") pod \"nmstate-operator-858ddd8f98-7l6n7\" (UID: \"4d407a70-cfdb-427d-94a8-c975db126733\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-7l6n7" Oct 13 08:55:26 crc kubenswrapper[4685]: I1013 08:55:26.997888 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfbj7\" (UniqueName: \"kubernetes.io/projected/4d407a70-cfdb-427d-94a8-c975db126733-kube-api-access-cfbj7\") pod \"nmstate-operator-858ddd8f98-7l6n7\" (UID: \"4d407a70-cfdb-427d-94a8-c975db126733\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-7l6n7" Oct 13 08:55:27 crc kubenswrapper[4685]: I1013 08:55:27.185345 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-7l6n7" Oct 13 08:55:27 crc kubenswrapper[4685]: I1013 08:55:27.395019 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-7l6n7"] Oct 13 08:55:28 crc kubenswrapper[4685]: I1013 08:55:28.168681 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-7l6n7" event={"ID":"4d407a70-cfdb-427d-94a8-c975db126733","Type":"ContainerStarted","Data":"9488d5b38333224dad0830b34df99d35b0a4c640d221722e5f69a89f7d4b4ceb"} Oct 13 08:55:30 crc kubenswrapper[4685]: I1013 08:55:30.181883 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-7l6n7" event={"ID":"4d407a70-cfdb-427d-94a8-c975db126733","Type":"ContainerStarted","Data":"633ce98d84b2f635bffb742b0ddfcb687c538d35d1e62e6ab99fc3df1ba90e9e"} Oct 13 08:55:35 crc kubenswrapper[4685]: I1013 08:55:35.907447 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-7l6n7" podStartSLOduration=7.828348418 podStartE2EDuration="9.907425756s" podCreationTimestamp="2025-10-13 08:55:26 +0000 UTC" firstStartedPulling="2025-10-13 08:55:27.403987483 +0000 UTC m=+652.551863244" lastFinishedPulling="2025-10-13 08:55:29.483064821 +0000 UTC m=+654.630940582" observedRunningTime="2025-10-13 08:55:30.209580872 +0000 UTC m=+655.357456643" watchObservedRunningTime="2025-10-13 08:55:35.907425756 +0000 UTC m=+661.055301507" Oct 13 08:55:35 crc kubenswrapper[4685]: I1013 08:55:35.908993 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-bgw9w"] Oct 13 08:55:35 crc kubenswrapper[4685]: I1013 08:55:35.910047 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bgw9w" Oct 13 08:55:35 crc kubenswrapper[4685]: I1013 08:55:35.911683 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-7b8xf" Oct 13 08:55:35 crc kubenswrapper[4685]: I1013 08:55:35.923593 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28th6\" (UniqueName: \"kubernetes.io/projected/4cdd7309-1689-4b3a-8165-bda0f92016c9-kube-api-access-28th6\") pod \"nmstate-metrics-fdff9cb8d-bgw9w\" (UID: \"4cdd7309-1689-4b3a-8165-bda0f92016c9\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bgw9w" Oct 13 08:55:35 crc kubenswrapper[4685]: I1013 08:55:35.925534 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-bgw9w"] Oct 13 08:55:35 crc kubenswrapper[4685]: I1013 08:55:35.962886 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-5cc6j"] Oct 13 08:55:35 crc kubenswrapper[4685]: I1013 08:55:35.963666 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-5cc6j" Oct 13 08:55:35 crc kubenswrapper[4685]: I1013 08:55:35.965234 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-66rzd"] Oct 13 08:55:35 crc kubenswrapper[4685]: I1013 08:55:35.965858 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-66rzd" Oct 13 08:55:35 crc kubenswrapper[4685]: I1013 08:55:35.969241 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 13 08:55:35 crc kubenswrapper[4685]: I1013 08:55:35.996635 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-66rzd"] Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.024508 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ed739611-c4b8-42d8-9c8f-3382d0347c39-nmstate-lock\") pod \"nmstate-handler-5cc6j\" (UID: \"ed739611-c4b8-42d8-9c8f-3382d0347c39\") " pod="openshift-nmstate/nmstate-handler-5cc6j" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.024640 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ed739611-c4b8-42d8-9c8f-3382d0347c39-ovs-socket\") pod \"nmstate-handler-5cc6j\" (UID: \"ed739611-c4b8-42d8-9c8f-3382d0347c39\") " pod="openshift-nmstate/nmstate-handler-5cc6j" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.024733 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ac20f0e3-6bff-43fa-93d5-6447bf249314-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-66rzd\" (UID: \"ac20f0e3-6bff-43fa-93d5-6447bf249314\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-66rzd" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.024951 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlm7b\" (UniqueName: \"kubernetes.io/projected/ac20f0e3-6bff-43fa-93d5-6447bf249314-kube-api-access-wlm7b\") pod \"nmstate-webhook-6cdbc54649-66rzd\" (UID: \"ac20f0e3-6bff-43fa-93d5-6447bf249314\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-66rzd" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.025103 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28th6\" (UniqueName: \"kubernetes.io/projected/4cdd7309-1689-4b3a-8165-bda0f92016c9-kube-api-access-28th6\") pod \"nmstate-metrics-fdff9cb8d-bgw9w\" (UID: \"4cdd7309-1689-4b3a-8165-bda0f92016c9\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bgw9w" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.025135 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ed739611-c4b8-42d8-9c8f-3382d0347c39-dbus-socket\") pod \"nmstate-handler-5cc6j\" (UID: \"ed739611-c4b8-42d8-9c8f-3382d0347c39\") " pod="openshift-nmstate/nmstate-handler-5cc6j" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.025156 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mr6mn\" (UniqueName: \"kubernetes.io/projected/ed739611-c4b8-42d8-9c8f-3382d0347c39-kube-api-access-mr6mn\") pod \"nmstate-handler-5cc6j\" (UID: \"ed739611-c4b8-42d8-9c8f-3382d0347c39\") " pod="openshift-nmstate/nmstate-handler-5cc6j" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.044759 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28th6\" (UniqueName: \"kubernetes.io/projected/4cdd7309-1689-4b3a-8165-bda0f92016c9-kube-api-access-28th6\") pod \"nmstate-metrics-fdff9cb8d-bgw9w\" (UID: \"4cdd7309-1689-4b3a-8165-bda0f92016c9\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bgw9w" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.089232 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-zscjx"] Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.089846 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-zscjx" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.092527 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.092835 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.093436 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-8tx4g" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.107437 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-zscjx"] Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.126814 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ed739611-c4b8-42d8-9c8f-3382d0347c39-dbus-socket\") pod \"nmstate-handler-5cc6j\" (UID: \"ed739611-c4b8-42d8-9c8f-3382d0347c39\") " pod="openshift-nmstate/nmstate-handler-5cc6j" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.126875 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mr6mn\" (UniqueName: \"kubernetes.io/projected/ed739611-c4b8-42d8-9c8f-3382d0347c39-kube-api-access-mr6mn\") pod \"nmstate-handler-5cc6j\" (UID: \"ed739611-c4b8-42d8-9c8f-3382d0347c39\") " pod="openshift-nmstate/nmstate-handler-5cc6j" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.126968 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1c0f8ea9-312a-479f-be32-33028d5d6651-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-zscjx\" (UID: \"1c0f8ea9-312a-479f-be32-33028d5d6651\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-zscjx" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.127000 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ed739611-c4b8-42d8-9c8f-3382d0347c39-nmstate-lock\") pod \"nmstate-handler-5cc6j\" (UID: \"ed739611-c4b8-42d8-9c8f-3382d0347c39\") " pod="openshift-nmstate/nmstate-handler-5cc6j" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.127026 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ed739611-c4b8-42d8-9c8f-3382d0347c39-ovs-socket\") pod \"nmstate-handler-5cc6j\" (UID: \"ed739611-c4b8-42d8-9c8f-3382d0347c39\") " pod="openshift-nmstate/nmstate-handler-5cc6j" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.127052 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ac20f0e3-6bff-43fa-93d5-6447bf249314-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-66rzd\" (UID: \"ac20f0e3-6bff-43fa-93d5-6447bf249314\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-66rzd" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.127086 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1c0f8ea9-312a-479f-be32-33028d5d6651-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-zscjx\" (UID: \"1c0f8ea9-312a-479f-be32-33028d5d6651\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-zscjx" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.127116 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlm7b\" (UniqueName: \"kubernetes.io/projected/ac20f0e3-6bff-43fa-93d5-6447bf249314-kube-api-access-wlm7b\") pod \"nmstate-webhook-6cdbc54649-66rzd\" (UID: \"ac20f0e3-6bff-43fa-93d5-6447bf249314\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-66rzd" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.127140 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f98qk\" (UniqueName: \"kubernetes.io/projected/1c0f8ea9-312a-479f-be32-33028d5d6651-kube-api-access-f98qk\") pod \"nmstate-console-plugin-6b874cbd85-zscjx\" (UID: \"1c0f8ea9-312a-479f-be32-33028d5d6651\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-zscjx" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.127219 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ed739611-c4b8-42d8-9c8f-3382d0347c39-dbus-socket\") pod \"nmstate-handler-5cc6j\" (UID: \"ed739611-c4b8-42d8-9c8f-3382d0347c39\") " pod="openshift-nmstate/nmstate-handler-5cc6j" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.127311 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ed739611-c4b8-42d8-9c8f-3382d0347c39-nmstate-lock\") pod \"nmstate-handler-5cc6j\" (UID: \"ed739611-c4b8-42d8-9c8f-3382d0347c39\") " pod="openshift-nmstate/nmstate-handler-5cc6j" Oct 13 08:55:36 crc kubenswrapper[4685]: E1013 08:55:36.127371 4685 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 13 08:55:36 crc kubenswrapper[4685]: E1013 08:55:36.127428 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ac20f0e3-6bff-43fa-93d5-6447bf249314-tls-key-pair podName:ac20f0e3-6bff-43fa-93d5-6447bf249314 nodeName:}" failed. No retries permitted until 2025-10-13 08:55:36.6274099 +0000 UTC m=+661.775285661 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/ac20f0e3-6bff-43fa-93d5-6447bf249314-tls-key-pair") pod "nmstate-webhook-6cdbc54649-66rzd" (UID: "ac20f0e3-6bff-43fa-93d5-6447bf249314") : secret "openshift-nmstate-webhook" not found Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.127510 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ed739611-c4b8-42d8-9c8f-3382d0347c39-ovs-socket\") pod \"nmstate-handler-5cc6j\" (UID: \"ed739611-c4b8-42d8-9c8f-3382d0347c39\") " pod="openshift-nmstate/nmstate-handler-5cc6j" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.143560 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlm7b\" (UniqueName: \"kubernetes.io/projected/ac20f0e3-6bff-43fa-93d5-6447bf249314-kube-api-access-wlm7b\") pod \"nmstate-webhook-6cdbc54649-66rzd\" (UID: \"ac20f0e3-6bff-43fa-93d5-6447bf249314\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-66rzd" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.145381 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mr6mn\" (UniqueName: \"kubernetes.io/projected/ed739611-c4b8-42d8-9c8f-3382d0347c39-kube-api-access-mr6mn\") pod \"nmstate-handler-5cc6j\" (UID: \"ed739611-c4b8-42d8-9c8f-3382d0347c39\") " pod="openshift-nmstate/nmstate-handler-5cc6j" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.225255 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bgw9w" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.228230 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1c0f8ea9-312a-479f-be32-33028d5d6651-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-zscjx\" (UID: \"1c0f8ea9-312a-479f-be32-33028d5d6651\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-zscjx" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.228390 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1c0f8ea9-312a-479f-be32-33028d5d6651-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-zscjx\" (UID: \"1c0f8ea9-312a-479f-be32-33028d5d6651\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-zscjx" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.228466 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f98qk\" (UniqueName: \"kubernetes.io/projected/1c0f8ea9-312a-479f-be32-33028d5d6651-kube-api-access-f98qk\") pod \"nmstate-console-plugin-6b874cbd85-zscjx\" (UID: \"1c0f8ea9-312a-479f-be32-33028d5d6651\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-zscjx" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.230012 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1c0f8ea9-312a-479f-be32-33028d5d6651-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-zscjx\" (UID: \"1c0f8ea9-312a-479f-be32-33028d5d6651\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-zscjx" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.232356 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1c0f8ea9-312a-479f-be32-33028d5d6651-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-zscjx\" (UID: \"1c0f8ea9-312a-479f-be32-33028d5d6651\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-zscjx" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.250633 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f98qk\" (UniqueName: \"kubernetes.io/projected/1c0f8ea9-312a-479f-be32-33028d5d6651-kube-api-access-f98qk\") pod \"nmstate-console-plugin-6b874cbd85-zscjx\" (UID: \"1c0f8ea9-312a-479f-be32-33028d5d6651\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-zscjx" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.278268 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-5cc6j" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.304009 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-54cc6bc5c6-kcdmf"] Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.304584 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: W1013 08:55:36.314378 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded739611_c4b8_42d8_9c8f_3382d0347c39.slice/crio-1d722b22e508d470ded7c689a8ca439c175754280131fd9e5bd7106176e37c69 WatchSource:0}: Error finding container 1d722b22e508d470ded7c689a8ca439c175754280131fd9e5bd7106176e37c69: Status 404 returned error can't find the container with id 1d722b22e508d470ded7c689a8ca439c175754280131fd9e5bd7106176e37c69 Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.329667 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/760b3555-c251-4b58-8a25-4be5f11ca1db-console-serving-cert\") pod \"console-54cc6bc5c6-kcdmf\" (UID: \"760b3555-c251-4b58-8a25-4be5f11ca1db\") " pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.329726 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/760b3555-c251-4b58-8a25-4be5f11ca1db-service-ca\") pod \"console-54cc6bc5c6-kcdmf\" (UID: \"760b3555-c251-4b58-8a25-4be5f11ca1db\") " pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.329750 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/760b3555-c251-4b58-8a25-4be5f11ca1db-console-oauth-config\") pod \"console-54cc6bc5c6-kcdmf\" (UID: \"760b3555-c251-4b58-8a25-4be5f11ca1db\") " pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.329792 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/760b3555-c251-4b58-8a25-4be5f11ca1db-trusted-ca-bundle\") pod \"console-54cc6bc5c6-kcdmf\" (UID: \"760b3555-c251-4b58-8a25-4be5f11ca1db\") " pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.329809 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/760b3555-c251-4b58-8a25-4be5f11ca1db-oauth-serving-cert\") pod \"console-54cc6bc5c6-kcdmf\" (UID: \"760b3555-c251-4b58-8a25-4be5f11ca1db\") " pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.329881 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/760b3555-c251-4b58-8a25-4be5f11ca1db-console-config\") pod \"console-54cc6bc5c6-kcdmf\" (UID: \"760b3555-c251-4b58-8a25-4be5f11ca1db\") " pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.329937 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gscsm\" (UniqueName: \"kubernetes.io/projected/760b3555-c251-4b58-8a25-4be5f11ca1db-kube-api-access-gscsm\") pod \"console-54cc6bc5c6-kcdmf\" (UID: \"760b3555-c251-4b58-8a25-4be5f11ca1db\") " pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.334572 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-54cc6bc5c6-kcdmf"] Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.402775 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-zscjx" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.435604 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/760b3555-c251-4b58-8a25-4be5f11ca1db-trusted-ca-bundle\") pod \"console-54cc6bc5c6-kcdmf\" (UID: \"760b3555-c251-4b58-8a25-4be5f11ca1db\") " pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.435952 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/760b3555-c251-4b58-8a25-4be5f11ca1db-oauth-serving-cert\") pod \"console-54cc6bc5c6-kcdmf\" (UID: \"760b3555-c251-4b58-8a25-4be5f11ca1db\") " pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.436014 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/760b3555-c251-4b58-8a25-4be5f11ca1db-console-config\") pod \"console-54cc6bc5c6-kcdmf\" (UID: \"760b3555-c251-4b58-8a25-4be5f11ca1db\") " pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.436049 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gscsm\" (UniqueName: \"kubernetes.io/projected/760b3555-c251-4b58-8a25-4be5f11ca1db-kube-api-access-gscsm\") pod \"console-54cc6bc5c6-kcdmf\" (UID: \"760b3555-c251-4b58-8a25-4be5f11ca1db\") " pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.436080 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/760b3555-c251-4b58-8a25-4be5f11ca1db-console-serving-cert\") pod \"console-54cc6bc5c6-kcdmf\" (UID: \"760b3555-c251-4b58-8a25-4be5f11ca1db\") " pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.436099 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/760b3555-c251-4b58-8a25-4be5f11ca1db-service-ca\") pod \"console-54cc6bc5c6-kcdmf\" (UID: \"760b3555-c251-4b58-8a25-4be5f11ca1db\") " pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.436118 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/760b3555-c251-4b58-8a25-4be5f11ca1db-console-oauth-config\") pod \"console-54cc6bc5c6-kcdmf\" (UID: \"760b3555-c251-4b58-8a25-4be5f11ca1db\") " pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.437006 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/760b3555-c251-4b58-8a25-4be5f11ca1db-trusted-ca-bundle\") pod \"console-54cc6bc5c6-kcdmf\" (UID: \"760b3555-c251-4b58-8a25-4be5f11ca1db\") " pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.438193 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/760b3555-c251-4b58-8a25-4be5f11ca1db-service-ca\") pod \"console-54cc6bc5c6-kcdmf\" (UID: \"760b3555-c251-4b58-8a25-4be5f11ca1db\") " pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.438766 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/760b3555-c251-4b58-8a25-4be5f11ca1db-oauth-serving-cert\") pod \"console-54cc6bc5c6-kcdmf\" (UID: \"760b3555-c251-4b58-8a25-4be5f11ca1db\") " pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.441701 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/760b3555-c251-4b58-8a25-4be5f11ca1db-console-oauth-config\") pod \"console-54cc6bc5c6-kcdmf\" (UID: \"760b3555-c251-4b58-8a25-4be5f11ca1db\") " pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.442137 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/760b3555-c251-4b58-8a25-4be5f11ca1db-console-config\") pod \"console-54cc6bc5c6-kcdmf\" (UID: \"760b3555-c251-4b58-8a25-4be5f11ca1db\") " pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.442605 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/760b3555-c251-4b58-8a25-4be5f11ca1db-console-serving-cert\") pod \"console-54cc6bc5c6-kcdmf\" (UID: \"760b3555-c251-4b58-8a25-4be5f11ca1db\") " pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.469869 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gscsm\" (UniqueName: \"kubernetes.io/projected/760b3555-c251-4b58-8a25-4be5f11ca1db-kube-api-access-gscsm\") pod \"console-54cc6bc5c6-kcdmf\" (UID: \"760b3555-c251-4b58-8a25-4be5f11ca1db\") " pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.622290 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-bgw9w"] Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.624306 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.640757 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ac20f0e3-6bff-43fa-93d5-6447bf249314-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-66rzd\" (UID: \"ac20f0e3-6bff-43fa-93d5-6447bf249314\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-66rzd" Oct 13 08:55:36 crc kubenswrapper[4685]: E1013 08:55:36.640946 4685 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 13 08:55:36 crc kubenswrapper[4685]: E1013 08:55:36.641004 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ac20f0e3-6bff-43fa-93d5-6447bf249314-tls-key-pair podName:ac20f0e3-6bff-43fa-93d5-6447bf249314 nodeName:}" failed. No retries permitted until 2025-10-13 08:55:37.640985831 +0000 UTC m=+662.788861592 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/ac20f0e3-6bff-43fa-93d5-6447bf249314-tls-key-pair") pod "nmstate-webhook-6cdbc54649-66rzd" (UID: "ac20f0e3-6bff-43fa-93d5-6447bf249314") : secret "openshift-nmstate-webhook" not found Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.688110 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-zscjx"] Oct 13 08:55:36 crc kubenswrapper[4685]: I1013 08:55:36.818034 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-54cc6bc5c6-kcdmf"] Oct 13 08:55:36 crc kubenswrapper[4685]: W1013 08:55:36.848805 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod760b3555_c251_4b58_8a25_4be5f11ca1db.slice/crio-1e0a773a266d789816b4a486f6baa5106cc344aba11b130d267524d4bead7f8d WatchSource:0}: Error finding container 1e0a773a266d789816b4a486f6baa5106cc344aba11b130d267524d4bead7f8d: Status 404 returned error can't find the container with id 1e0a773a266d789816b4a486f6baa5106cc344aba11b130d267524d4bead7f8d Oct 13 08:55:37 crc kubenswrapper[4685]: I1013 08:55:37.224995 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-zscjx" event={"ID":"1c0f8ea9-312a-479f-be32-33028d5d6651","Type":"ContainerStarted","Data":"24312be782245f208521fad178b3669af0fdbcdd187d7af3a4f1acf9e3618982"} Oct 13 08:55:37 crc kubenswrapper[4685]: I1013 08:55:37.228547 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bgw9w" event={"ID":"4cdd7309-1689-4b3a-8165-bda0f92016c9","Type":"ContainerStarted","Data":"cc94a5ac2769f25437e9f29b1e5a912154eecb521f3b72f055c64e6d3eb57c16"} Oct 13 08:55:37 crc kubenswrapper[4685]: I1013 08:55:37.230091 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-5cc6j" event={"ID":"ed739611-c4b8-42d8-9c8f-3382d0347c39","Type":"ContainerStarted","Data":"1d722b22e508d470ded7c689a8ca439c175754280131fd9e5bd7106176e37c69"} Oct 13 08:55:37 crc kubenswrapper[4685]: I1013 08:55:37.232990 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-54cc6bc5c6-kcdmf" event={"ID":"760b3555-c251-4b58-8a25-4be5f11ca1db","Type":"ContainerStarted","Data":"229009d5a6cebd3fa8bb39dcfa77be803feb7ab37837c278770a503ebcdb4911"} Oct 13 08:55:37 crc kubenswrapper[4685]: I1013 08:55:37.233061 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-54cc6bc5c6-kcdmf" event={"ID":"760b3555-c251-4b58-8a25-4be5f11ca1db","Type":"ContainerStarted","Data":"1e0a773a266d789816b4a486f6baa5106cc344aba11b130d267524d4bead7f8d"} Oct 13 08:55:37 crc kubenswrapper[4685]: I1013 08:55:37.268359 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-54cc6bc5c6-kcdmf" podStartSLOduration=1.2683317139999999 podStartE2EDuration="1.268331714s" podCreationTimestamp="2025-10-13 08:55:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:55:37.265225408 +0000 UTC m=+662.413101259" watchObservedRunningTime="2025-10-13 08:55:37.268331714 +0000 UTC m=+662.416207505" Oct 13 08:55:37 crc kubenswrapper[4685]: I1013 08:55:37.659725 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ac20f0e3-6bff-43fa-93d5-6447bf249314-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-66rzd\" (UID: \"ac20f0e3-6bff-43fa-93d5-6447bf249314\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-66rzd" Oct 13 08:55:37 crc kubenswrapper[4685]: I1013 08:55:37.683833 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ac20f0e3-6bff-43fa-93d5-6447bf249314-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-66rzd\" (UID: \"ac20f0e3-6bff-43fa-93d5-6447bf249314\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-66rzd" Oct 13 08:55:37 crc kubenswrapper[4685]: I1013 08:55:37.783903 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-66rzd" Oct 13 08:55:38 crc kubenswrapper[4685]: I1013 08:55:38.028378 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-66rzd"] Oct 13 08:55:38 crc kubenswrapper[4685]: I1013 08:55:38.241658 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-66rzd" event={"ID":"ac20f0e3-6bff-43fa-93d5-6447bf249314","Type":"ContainerStarted","Data":"4ff478c17bfbe37f8d95db59b28e1b61472cbb93ba8a6cde69dbfe063f25c534"} Oct 13 08:55:40 crc kubenswrapper[4685]: I1013 08:55:40.257446 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-66rzd" event={"ID":"ac20f0e3-6bff-43fa-93d5-6447bf249314","Type":"ContainerStarted","Data":"74c897ac06110f29644235afeb13d344f8cc8d0a8b36a533393fed7d41ddaaba"} Oct 13 08:55:40 crc kubenswrapper[4685]: I1013 08:55:40.258219 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-66rzd" Oct 13 08:55:40 crc kubenswrapper[4685]: I1013 08:55:40.260373 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-5cc6j" event={"ID":"ed739611-c4b8-42d8-9c8f-3382d0347c39","Type":"ContainerStarted","Data":"43f7efb9350eb9e5e9ffcc56978a28dd8adf2788b301f69397afe7aca1db6404"} Oct 13 08:55:40 crc kubenswrapper[4685]: I1013 08:55:40.260445 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-5cc6j" Oct 13 08:55:40 crc kubenswrapper[4685]: I1013 08:55:40.264055 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-zscjx" event={"ID":"1c0f8ea9-312a-479f-be32-33028d5d6651","Type":"ContainerStarted","Data":"0cedc53d42ba3f8cabb1993e339df8ace2aa3a4e18cae7aec3fcc57f04547dc1"} Oct 13 08:55:40 crc kubenswrapper[4685]: I1013 08:55:40.267413 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bgw9w" event={"ID":"4cdd7309-1689-4b3a-8165-bda0f92016c9","Type":"ContainerStarted","Data":"4d763d7cd43535ab2c69d78449be75f1ecb729f79115cdce5c62fc0f70d6adcb"} Oct 13 08:55:40 crc kubenswrapper[4685]: I1013 08:55:40.279338 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-66rzd" podStartSLOduration=4.044614701 podStartE2EDuration="5.279315776s" podCreationTimestamp="2025-10-13 08:55:35 +0000 UTC" firstStartedPulling="2025-10-13 08:55:38.037393265 +0000 UTC m=+663.185269026" lastFinishedPulling="2025-10-13 08:55:39.27209434 +0000 UTC m=+664.419970101" observedRunningTime="2025-10-13 08:55:40.275896242 +0000 UTC m=+665.423772013" watchObservedRunningTime="2025-10-13 08:55:40.279315776 +0000 UTC m=+665.427191547" Oct 13 08:55:40 crc kubenswrapper[4685]: I1013 08:55:40.322644 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-5cc6j" podStartSLOduration=2.343695948 podStartE2EDuration="5.322623477s" podCreationTimestamp="2025-10-13 08:55:35 +0000 UTC" firstStartedPulling="2025-10-13 08:55:36.324449512 +0000 UTC m=+661.472325273" lastFinishedPulling="2025-10-13 08:55:39.303377031 +0000 UTC m=+664.451252802" observedRunningTime="2025-10-13 08:55:40.320761026 +0000 UTC m=+665.468636787" watchObservedRunningTime="2025-10-13 08:55:40.322623477 +0000 UTC m=+665.470499248" Oct 13 08:55:40 crc kubenswrapper[4685]: I1013 08:55:40.323996 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-zscjx" podStartSLOduration=1.755212821 podStartE2EDuration="4.323991105s" podCreationTimestamp="2025-10-13 08:55:36 +0000 UTC" firstStartedPulling="2025-10-13 08:55:36.700322084 +0000 UTC m=+661.848197845" lastFinishedPulling="2025-10-13 08:55:39.269100368 +0000 UTC m=+664.416976129" observedRunningTime="2025-10-13 08:55:40.300023965 +0000 UTC m=+665.447899736" watchObservedRunningTime="2025-10-13 08:55:40.323991105 +0000 UTC m=+665.471866876" Oct 13 08:55:42 crc kubenswrapper[4685]: I1013 08:55:42.284791 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bgw9w" event={"ID":"4cdd7309-1689-4b3a-8165-bda0f92016c9","Type":"ContainerStarted","Data":"df29a84c7f8a406994e93e51949ed59ba1afd736dc83d07a3842dc8fc088a500"} Oct 13 08:55:42 crc kubenswrapper[4685]: I1013 08:55:42.315087 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bgw9w" podStartSLOduration=2.199776277 podStartE2EDuration="7.315031911s" podCreationTimestamp="2025-10-13 08:55:35 +0000 UTC" firstStartedPulling="2025-10-13 08:55:36.636525468 +0000 UTC m=+661.784401229" lastFinishedPulling="2025-10-13 08:55:41.751781112 +0000 UTC m=+666.899656863" observedRunningTime="2025-10-13 08:55:42.308750458 +0000 UTC m=+667.456626269" watchObservedRunningTime="2025-10-13 08:55:42.315031911 +0000 UTC m=+667.462907712" Oct 13 08:55:46 crc kubenswrapper[4685]: I1013 08:55:46.312965 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-5cc6j" Oct 13 08:55:46 crc kubenswrapper[4685]: I1013 08:55:46.625640 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:46 crc kubenswrapper[4685]: I1013 08:55:46.626103 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:46 crc kubenswrapper[4685]: I1013 08:55:46.633973 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:47 crc kubenswrapper[4685]: I1013 08:55:47.323320 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-54cc6bc5c6-kcdmf" Oct 13 08:55:47 crc kubenswrapper[4685]: I1013 08:55:47.398797 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-gx6qj"] Oct 13 08:55:57 crc kubenswrapper[4685]: I1013 08:55:57.794030 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-66rzd" Oct 13 08:56:11 crc kubenswrapper[4685]: I1013 08:56:11.995253 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z"] Oct 13 08:56:11 crc kubenswrapper[4685]: I1013 08:56:11.998365 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z" Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.001215 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.010505 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z"] Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.041961 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsgbv\" (UniqueName: \"kubernetes.io/projected/624e992c-aa90-41ff-a746-adf45015b81a-kube-api-access-tsgbv\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z\" (UID: \"624e992c-aa90-41ff-a746-adf45015b81a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z" Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.042299 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/624e992c-aa90-41ff-a746-adf45015b81a-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z\" (UID: \"624e992c-aa90-41ff-a746-adf45015b81a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z" Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.042509 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/624e992c-aa90-41ff-a746-adf45015b81a-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z\" (UID: \"624e992c-aa90-41ff-a746-adf45015b81a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z" Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.144049 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/624e992c-aa90-41ff-a746-adf45015b81a-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z\" (UID: \"624e992c-aa90-41ff-a746-adf45015b81a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z" Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.144178 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsgbv\" (UniqueName: \"kubernetes.io/projected/624e992c-aa90-41ff-a746-adf45015b81a-kube-api-access-tsgbv\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z\" (UID: \"624e992c-aa90-41ff-a746-adf45015b81a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z" Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.144216 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/624e992c-aa90-41ff-a746-adf45015b81a-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z\" (UID: \"624e992c-aa90-41ff-a746-adf45015b81a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z" Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.144890 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/624e992c-aa90-41ff-a746-adf45015b81a-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z\" (UID: \"624e992c-aa90-41ff-a746-adf45015b81a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z" Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.145538 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/624e992c-aa90-41ff-a746-adf45015b81a-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z\" (UID: \"624e992c-aa90-41ff-a746-adf45015b81a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z" Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.166270 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsgbv\" (UniqueName: \"kubernetes.io/projected/624e992c-aa90-41ff-a746-adf45015b81a-kube-api-access-tsgbv\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z\" (UID: \"624e992c-aa90-41ff-a746-adf45015b81a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z" Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.326412 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z" Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.472300 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-gx6qj" podUID="16484304-6976-45e8-a495-f1d2ee367f0f" containerName="console" containerID="cri-o://eb84cf90b0898dd2fc99fab0c114053a787806daf85967a3410211428a74a350" gracePeriod=15 Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.673687 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z"] Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.918164 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-gx6qj_16484304-6976-45e8-a495-f1d2ee367f0f/console/0.log" Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.918463 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.955601 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-trusted-ca-bundle\") pod \"16484304-6976-45e8-a495-f1d2ee367f0f\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.955680 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-service-ca\") pod \"16484304-6976-45e8-a495-f1d2ee367f0f\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.955712 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-console-config\") pod \"16484304-6976-45e8-a495-f1d2ee367f0f\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.955776 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/16484304-6976-45e8-a495-f1d2ee367f0f-console-serving-cert\") pod \"16484304-6976-45e8-a495-f1d2ee367f0f\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.955804 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hfd8\" (UniqueName: \"kubernetes.io/projected/16484304-6976-45e8-a495-f1d2ee367f0f-kube-api-access-6hfd8\") pod \"16484304-6976-45e8-a495-f1d2ee367f0f\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.955863 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/16484304-6976-45e8-a495-f1d2ee367f0f-console-oauth-config\") pod \"16484304-6976-45e8-a495-f1d2ee367f0f\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.955884 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-oauth-serving-cert\") pod \"16484304-6976-45e8-a495-f1d2ee367f0f\" (UID: \"16484304-6976-45e8-a495-f1d2ee367f0f\") " Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.956604 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "16484304-6976-45e8-a495-f1d2ee367f0f" (UID: "16484304-6976-45e8-a495-f1d2ee367f0f"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.956851 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-service-ca" (OuterVolumeSpecName: "service-ca") pod "16484304-6976-45e8-a495-f1d2ee367f0f" (UID: "16484304-6976-45e8-a495-f1d2ee367f0f"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.957131 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "16484304-6976-45e8-a495-f1d2ee367f0f" (UID: "16484304-6976-45e8-a495-f1d2ee367f0f"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.957124 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-console-config" (OuterVolumeSpecName: "console-config") pod "16484304-6976-45e8-a495-f1d2ee367f0f" (UID: "16484304-6976-45e8-a495-f1d2ee367f0f"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.961114 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16484304-6976-45e8-a495-f1d2ee367f0f-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "16484304-6976-45e8-a495-f1d2ee367f0f" (UID: "16484304-6976-45e8-a495-f1d2ee367f0f"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.961465 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16484304-6976-45e8-a495-f1d2ee367f0f-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "16484304-6976-45e8-a495-f1d2ee367f0f" (UID: "16484304-6976-45e8-a495-f1d2ee367f0f"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:56:12 crc kubenswrapper[4685]: I1013 08:56:12.961699 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16484304-6976-45e8-a495-f1d2ee367f0f-kube-api-access-6hfd8" (OuterVolumeSpecName: "kube-api-access-6hfd8") pod "16484304-6976-45e8-a495-f1d2ee367f0f" (UID: "16484304-6976-45e8-a495-f1d2ee367f0f"). InnerVolumeSpecName "kube-api-access-6hfd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:56:13 crc kubenswrapper[4685]: I1013 08:56:13.057384 4685 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 08:56:13 crc kubenswrapper[4685]: I1013 08:56:13.057414 4685 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-service-ca\") on node \"crc\" DevicePath \"\"" Oct 13 08:56:13 crc kubenswrapper[4685]: I1013 08:56:13.057423 4685 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-console-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:56:13 crc kubenswrapper[4685]: I1013 08:56:13.057433 4685 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/16484304-6976-45e8-a495-f1d2ee367f0f-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:56:13 crc kubenswrapper[4685]: I1013 08:56:13.057441 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hfd8\" (UniqueName: \"kubernetes.io/projected/16484304-6976-45e8-a495-f1d2ee367f0f-kube-api-access-6hfd8\") on node \"crc\" DevicePath \"\"" Oct 13 08:56:13 crc kubenswrapper[4685]: I1013 08:56:13.057450 4685 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/16484304-6976-45e8-a495-f1d2ee367f0f-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:56:13 crc kubenswrapper[4685]: I1013 08:56:13.057458 4685 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/16484304-6976-45e8-a495-f1d2ee367f0f-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:56:13 crc kubenswrapper[4685]: I1013 08:56:13.493175 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-gx6qj_16484304-6976-45e8-a495-f1d2ee367f0f/console/0.log" Oct 13 08:56:13 crc kubenswrapper[4685]: I1013 08:56:13.493260 4685 generic.go:334] "Generic (PLEG): container finished" podID="16484304-6976-45e8-a495-f1d2ee367f0f" containerID="eb84cf90b0898dd2fc99fab0c114053a787806daf85967a3410211428a74a350" exitCode=2 Oct 13 08:56:13 crc kubenswrapper[4685]: I1013 08:56:13.493378 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-gx6qj" event={"ID":"16484304-6976-45e8-a495-f1d2ee367f0f","Type":"ContainerDied","Data":"eb84cf90b0898dd2fc99fab0c114053a787806daf85967a3410211428a74a350"} Oct 13 08:56:13 crc kubenswrapper[4685]: I1013 08:56:13.493385 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-gx6qj" Oct 13 08:56:13 crc kubenswrapper[4685]: I1013 08:56:13.493418 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-gx6qj" event={"ID":"16484304-6976-45e8-a495-f1d2ee367f0f","Type":"ContainerDied","Data":"46405a1c5e6d6de272837e94e5d775db5fbb4d672663784fb24a92c82bb93021"} Oct 13 08:56:13 crc kubenswrapper[4685]: I1013 08:56:13.493446 4685 scope.go:117] "RemoveContainer" containerID="eb84cf90b0898dd2fc99fab0c114053a787806daf85967a3410211428a74a350" Oct 13 08:56:13 crc kubenswrapper[4685]: I1013 08:56:13.498142 4685 generic.go:334] "Generic (PLEG): container finished" podID="624e992c-aa90-41ff-a746-adf45015b81a" containerID="681e98633f920c0944cb58f4d7cdba3f5b97f213580e6d98773c097d7e0260c9" exitCode=0 Oct 13 08:56:13 crc kubenswrapper[4685]: I1013 08:56:13.498212 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z" event={"ID":"624e992c-aa90-41ff-a746-adf45015b81a","Type":"ContainerDied","Data":"681e98633f920c0944cb58f4d7cdba3f5b97f213580e6d98773c097d7e0260c9"} Oct 13 08:56:13 crc kubenswrapper[4685]: I1013 08:56:13.498249 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z" event={"ID":"624e992c-aa90-41ff-a746-adf45015b81a","Type":"ContainerStarted","Data":"dd2e624cff595301374122505eab243e5bc4f6f151e39428673e26e38f6bc9e3"} Oct 13 08:56:13 crc kubenswrapper[4685]: I1013 08:56:13.526636 4685 scope.go:117] "RemoveContainer" containerID="eb84cf90b0898dd2fc99fab0c114053a787806daf85967a3410211428a74a350" Oct 13 08:56:13 crc kubenswrapper[4685]: E1013 08:56:13.527582 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb84cf90b0898dd2fc99fab0c114053a787806daf85967a3410211428a74a350\": container with ID starting with eb84cf90b0898dd2fc99fab0c114053a787806daf85967a3410211428a74a350 not found: ID does not exist" containerID="eb84cf90b0898dd2fc99fab0c114053a787806daf85967a3410211428a74a350" Oct 13 08:56:13 crc kubenswrapper[4685]: I1013 08:56:13.527809 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb84cf90b0898dd2fc99fab0c114053a787806daf85967a3410211428a74a350"} err="failed to get container status \"eb84cf90b0898dd2fc99fab0c114053a787806daf85967a3410211428a74a350\": rpc error: code = NotFound desc = could not find container \"eb84cf90b0898dd2fc99fab0c114053a787806daf85967a3410211428a74a350\": container with ID starting with eb84cf90b0898dd2fc99fab0c114053a787806daf85967a3410211428a74a350 not found: ID does not exist" Oct 13 08:56:13 crc kubenswrapper[4685]: I1013 08:56:13.560897 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-gx6qj"] Oct 13 08:56:13 crc kubenswrapper[4685]: I1013 08:56:13.570640 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-gx6qj"] Oct 13 08:56:15 crc kubenswrapper[4685]: I1013 08:56:15.519059 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16484304-6976-45e8-a495-f1d2ee367f0f" path="/var/lib/kubelet/pods/16484304-6976-45e8-a495-f1d2ee367f0f/volumes" Oct 13 08:56:15 crc kubenswrapper[4685]: I1013 08:56:15.521007 4685 generic.go:334] "Generic (PLEG): container finished" podID="624e992c-aa90-41ff-a746-adf45015b81a" containerID="3e2df282869b57d12ac6e79a0fe1daaac6bcb2516ca3b4f5e67858de85cda309" exitCode=0 Oct 13 08:56:15 crc kubenswrapper[4685]: I1013 08:56:15.521053 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z" event={"ID":"624e992c-aa90-41ff-a746-adf45015b81a","Type":"ContainerDied","Data":"3e2df282869b57d12ac6e79a0fe1daaac6bcb2516ca3b4f5e67858de85cda309"} Oct 13 08:56:16 crc kubenswrapper[4685]: I1013 08:56:16.533568 4685 generic.go:334] "Generic (PLEG): container finished" podID="624e992c-aa90-41ff-a746-adf45015b81a" containerID="2949aa1cd52c5f2c710e51da5ce74b043344d80bd452d05c34cdf6f8169522fa" exitCode=0 Oct 13 08:56:16 crc kubenswrapper[4685]: I1013 08:56:16.533657 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z" event={"ID":"624e992c-aa90-41ff-a746-adf45015b81a","Type":"ContainerDied","Data":"2949aa1cd52c5f2c710e51da5ce74b043344d80bd452d05c34cdf6f8169522fa"} Oct 13 08:56:17 crc kubenswrapper[4685]: I1013 08:56:17.814765 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z" Oct 13 08:56:17 crc kubenswrapper[4685]: I1013 08:56:17.923406 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsgbv\" (UniqueName: \"kubernetes.io/projected/624e992c-aa90-41ff-a746-adf45015b81a-kube-api-access-tsgbv\") pod \"624e992c-aa90-41ff-a746-adf45015b81a\" (UID: \"624e992c-aa90-41ff-a746-adf45015b81a\") " Oct 13 08:56:17 crc kubenswrapper[4685]: I1013 08:56:17.923479 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/624e992c-aa90-41ff-a746-adf45015b81a-bundle\") pod \"624e992c-aa90-41ff-a746-adf45015b81a\" (UID: \"624e992c-aa90-41ff-a746-adf45015b81a\") " Oct 13 08:56:17 crc kubenswrapper[4685]: I1013 08:56:17.923530 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/624e992c-aa90-41ff-a746-adf45015b81a-util\") pod \"624e992c-aa90-41ff-a746-adf45015b81a\" (UID: \"624e992c-aa90-41ff-a746-adf45015b81a\") " Oct 13 08:56:17 crc kubenswrapper[4685]: I1013 08:56:17.925645 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/624e992c-aa90-41ff-a746-adf45015b81a-bundle" (OuterVolumeSpecName: "bundle") pod "624e992c-aa90-41ff-a746-adf45015b81a" (UID: "624e992c-aa90-41ff-a746-adf45015b81a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:56:17 crc kubenswrapper[4685]: I1013 08:56:17.931845 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/624e992c-aa90-41ff-a746-adf45015b81a-kube-api-access-tsgbv" (OuterVolumeSpecName: "kube-api-access-tsgbv") pod "624e992c-aa90-41ff-a746-adf45015b81a" (UID: "624e992c-aa90-41ff-a746-adf45015b81a"). InnerVolumeSpecName "kube-api-access-tsgbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:56:17 crc kubenswrapper[4685]: I1013 08:56:17.955335 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/624e992c-aa90-41ff-a746-adf45015b81a-util" (OuterVolumeSpecName: "util") pod "624e992c-aa90-41ff-a746-adf45015b81a" (UID: "624e992c-aa90-41ff-a746-adf45015b81a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:56:18 crc kubenswrapper[4685]: I1013 08:56:18.024696 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsgbv\" (UniqueName: \"kubernetes.io/projected/624e992c-aa90-41ff-a746-adf45015b81a-kube-api-access-tsgbv\") on node \"crc\" DevicePath \"\"" Oct 13 08:56:18 crc kubenswrapper[4685]: I1013 08:56:18.024731 4685 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/624e992c-aa90-41ff-a746-adf45015b81a-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 08:56:18 crc kubenswrapper[4685]: I1013 08:56:18.024743 4685 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/624e992c-aa90-41ff-a746-adf45015b81a-util\") on node \"crc\" DevicePath \"\"" Oct 13 08:56:18 crc kubenswrapper[4685]: I1013 08:56:18.552700 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z" event={"ID":"624e992c-aa90-41ff-a746-adf45015b81a","Type":"ContainerDied","Data":"dd2e624cff595301374122505eab243e5bc4f6f151e39428673e26e38f6bc9e3"} Oct 13 08:56:18 crc kubenswrapper[4685]: I1013 08:56:18.552775 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd2e624cff595301374122505eab243e5bc4f6f151e39428673e26e38f6bc9e3" Oct 13 08:56:18 crc kubenswrapper[4685]: I1013 08:56:18.552801 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z" Oct 13 08:56:26 crc kubenswrapper[4685]: I1013 08:56:26.968667 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-759fc95d5d-79ckz"] Oct 13 08:56:26 crc kubenswrapper[4685]: E1013 08:56:26.969466 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="624e992c-aa90-41ff-a746-adf45015b81a" containerName="pull" Oct 13 08:56:26 crc kubenswrapper[4685]: I1013 08:56:26.969484 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="624e992c-aa90-41ff-a746-adf45015b81a" containerName="pull" Oct 13 08:56:26 crc kubenswrapper[4685]: E1013 08:56:26.969507 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="624e992c-aa90-41ff-a746-adf45015b81a" containerName="util" Oct 13 08:56:26 crc kubenswrapper[4685]: I1013 08:56:26.969515 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="624e992c-aa90-41ff-a746-adf45015b81a" containerName="util" Oct 13 08:56:26 crc kubenswrapper[4685]: E1013 08:56:26.969526 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="624e992c-aa90-41ff-a746-adf45015b81a" containerName="extract" Oct 13 08:56:26 crc kubenswrapper[4685]: I1013 08:56:26.969534 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="624e992c-aa90-41ff-a746-adf45015b81a" containerName="extract" Oct 13 08:56:26 crc kubenswrapper[4685]: E1013 08:56:26.969548 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16484304-6976-45e8-a495-f1d2ee367f0f" containerName="console" Oct 13 08:56:26 crc kubenswrapper[4685]: I1013 08:56:26.969555 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="16484304-6976-45e8-a495-f1d2ee367f0f" containerName="console" Oct 13 08:56:26 crc kubenswrapper[4685]: I1013 08:56:26.969672 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="624e992c-aa90-41ff-a746-adf45015b81a" containerName="extract" Oct 13 08:56:26 crc kubenswrapper[4685]: I1013 08:56:26.969688 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="16484304-6976-45e8-a495-f1d2ee367f0f" containerName="console" Oct 13 08:56:26 crc kubenswrapper[4685]: I1013 08:56:26.970084 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-759fc95d5d-79ckz" Oct 13 08:56:26 crc kubenswrapper[4685]: I1013 08:56:26.974792 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 13 08:56:26 crc kubenswrapper[4685]: I1013 08:56:26.974821 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-qzpjb" Oct 13 08:56:26 crc kubenswrapper[4685]: I1013 08:56:26.977482 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 13 08:56:26 crc kubenswrapper[4685]: I1013 08:56:26.980900 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 13 08:56:26 crc kubenswrapper[4685]: I1013 08:56:26.985980 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.033715 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-759fc95d5d-79ckz"] Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.042868 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gpg6\" (UniqueName: \"kubernetes.io/projected/b3d2fe67-48db-4dbc-869d-d0c13b18ec8a-kube-api-access-4gpg6\") pod \"metallb-operator-controller-manager-759fc95d5d-79ckz\" (UID: \"b3d2fe67-48db-4dbc-869d-d0c13b18ec8a\") " pod="metallb-system/metallb-operator-controller-manager-759fc95d5d-79ckz" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.042946 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b3d2fe67-48db-4dbc-869d-d0c13b18ec8a-webhook-cert\") pod \"metallb-operator-controller-manager-759fc95d5d-79ckz\" (UID: \"b3d2fe67-48db-4dbc-869d-d0c13b18ec8a\") " pod="metallb-system/metallb-operator-controller-manager-759fc95d5d-79ckz" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.043000 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b3d2fe67-48db-4dbc-869d-d0c13b18ec8a-apiservice-cert\") pod \"metallb-operator-controller-manager-759fc95d5d-79ckz\" (UID: \"b3d2fe67-48db-4dbc-869d-d0c13b18ec8a\") " pod="metallb-system/metallb-operator-controller-manager-759fc95d5d-79ckz" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.144182 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b3d2fe67-48db-4dbc-869d-d0c13b18ec8a-apiservice-cert\") pod \"metallb-operator-controller-manager-759fc95d5d-79ckz\" (UID: \"b3d2fe67-48db-4dbc-869d-d0c13b18ec8a\") " pod="metallb-system/metallb-operator-controller-manager-759fc95d5d-79ckz" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.144262 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gpg6\" (UniqueName: \"kubernetes.io/projected/b3d2fe67-48db-4dbc-869d-d0c13b18ec8a-kube-api-access-4gpg6\") pod \"metallb-operator-controller-manager-759fc95d5d-79ckz\" (UID: \"b3d2fe67-48db-4dbc-869d-d0c13b18ec8a\") " pod="metallb-system/metallb-operator-controller-manager-759fc95d5d-79ckz" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.144324 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b3d2fe67-48db-4dbc-869d-d0c13b18ec8a-webhook-cert\") pod \"metallb-operator-controller-manager-759fc95d5d-79ckz\" (UID: \"b3d2fe67-48db-4dbc-869d-d0c13b18ec8a\") " pod="metallb-system/metallb-operator-controller-manager-759fc95d5d-79ckz" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.152675 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b3d2fe67-48db-4dbc-869d-d0c13b18ec8a-webhook-cert\") pod \"metallb-operator-controller-manager-759fc95d5d-79ckz\" (UID: \"b3d2fe67-48db-4dbc-869d-d0c13b18ec8a\") " pod="metallb-system/metallb-operator-controller-manager-759fc95d5d-79ckz" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.181718 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b3d2fe67-48db-4dbc-869d-d0c13b18ec8a-apiservice-cert\") pod \"metallb-operator-controller-manager-759fc95d5d-79ckz\" (UID: \"b3d2fe67-48db-4dbc-869d-d0c13b18ec8a\") " pod="metallb-system/metallb-operator-controller-manager-759fc95d5d-79ckz" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.182571 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gpg6\" (UniqueName: \"kubernetes.io/projected/b3d2fe67-48db-4dbc-869d-d0c13b18ec8a-kube-api-access-4gpg6\") pod \"metallb-operator-controller-manager-759fc95d5d-79ckz\" (UID: \"b3d2fe67-48db-4dbc-869d-d0c13b18ec8a\") " pod="metallb-system/metallb-operator-controller-manager-759fc95d5d-79ckz" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.283883 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-759fc95d5d-79ckz" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.326104 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6f5fffbd86-pw7w6"] Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.326837 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6f5fffbd86-pw7w6" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.330698 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.331020 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-4vsbl" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.331167 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.347561 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6f5fffbd86-pw7w6"] Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.448819 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ce259f59-820d-41e4-8ef7-3b977664b7ea-webhook-cert\") pod \"metallb-operator-webhook-server-6f5fffbd86-pw7w6\" (UID: \"ce259f59-820d-41e4-8ef7-3b977664b7ea\") " pod="metallb-system/metallb-operator-webhook-server-6f5fffbd86-pw7w6" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.448964 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ce259f59-820d-41e4-8ef7-3b977664b7ea-apiservice-cert\") pod \"metallb-operator-webhook-server-6f5fffbd86-pw7w6\" (UID: \"ce259f59-820d-41e4-8ef7-3b977664b7ea\") " pod="metallb-system/metallb-operator-webhook-server-6f5fffbd86-pw7w6" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.449022 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jnl2\" (UniqueName: \"kubernetes.io/projected/ce259f59-820d-41e4-8ef7-3b977664b7ea-kube-api-access-8jnl2\") pod \"metallb-operator-webhook-server-6f5fffbd86-pw7w6\" (UID: \"ce259f59-820d-41e4-8ef7-3b977664b7ea\") " pod="metallb-system/metallb-operator-webhook-server-6f5fffbd86-pw7w6" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.554288 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ce259f59-820d-41e4-8ef7-3b977664b7ea-webhook-cert\") pod \"metallb-operator-webhook-server-6f5fffbd86-pw7w6\" (UID: \"ce259f59-820d-41e4-8ef7-3b977664b7ea\") " pod="metallb-system/metallb-operator-webhook-server-6f5fffbd86-pw7w6" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.554366 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ce259f59-820d-41e4-8ef7-3b977664b7ea-apiservice-cert\") pod \"metallb-operator-webhook-server-6f5fffbd86-pw7w6\" (UID: \"ce259f59-820d-41e4-8ef7-3b977664b7ea\") " pod="metallb-system/metallb-operator-webhook-server-6f5fffbd86-pw7w6" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.554395 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jnl2\" (UniqueName: \"kubernetes.io/projected/ce259f59-820d-41e4-8ef7-3b977664b7ea-kube-api-access-8jnl2\") pod \"metallb-operator-webhook-server-6f5fffbd86-pw7w6\" (UID: \"ce259f59-820d-41e4-8ef7-3b977664b7ea\") " pod="metallb-system/metallb-operator-webhook-server-6f5fffbd86-pw7w6" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.582361 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jnl2\" (UniqueName: \"kubernetes.io/projected/ce259f59-820d-41e4-8ef7-3b977664b7ea-kube-api-access-8jnl2\") pod \"metallb-operator-webhook-server-6f5fffbd86-pw7w6\" (UID: \"ce259f59-820d-41e4-8ef7-3b977664b7ea\") " pod="metallb-system/metallb-operator-webhook-server-6f5fffbd86-pw7w6" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.583688 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-759fc95d5d-79ckz"] Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.587376 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ce259f59-820d-41e4-8ef7-3b977664b7ea-webhook-cert\") pod \"metallb-operator-webhook-server-6f5fffbd86-pw7w6\" (UID: \"ce259f59-820d-41e4-8ef7-3b977664b7ea\") " pod="metallb-system/metallb-operator-webhook-server-6f5fffbd86-pw7w6" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.590579 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ce259f59-820d-41e4-8ef7-3b977664b7ea-apiservice-cert\") pod \"metallb-operator-webhook-server-6f5fffbd86-pw7w6\" (UID: \"ce259f59-820d-41e4-8ef7-3b977664b7ea\") " pod="metallb-system/metallb-operator-webhook-server-6f5fffbd86-pw7w6" Oct 13 08:56:27 crc kubenswrapper[4685]: W1013 08:56:27.603031 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3d2fe67_48db_4dbc_869d_d0c13b18ec8a.slice/crio-bd5b6bca22073a4ad518262465916489fc343c2a2b8aecf62c8b9bd02b95bac5 WatchSource:0}: Error finding container bd5b6bca22073a4ad518262465916489fc343c2a2b8aecf62c8b9bd02b95bac5: Status 404 returned error can't find the container with id bd5b6bca22073a4ad518262465916489fc343c2a2b8aecf62c8b9bd02b95bac5 Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.667282 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6f5fffbd86-pw7w6" Oct 13 08:56:27 crc kubenswrapper[4685]: I1013 08:56:27.907867 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6f5fffbd86-pw7w6"] Oct 13 08:56:27 crc kubenswrapper[4685]: W1013 08:56:27.915172 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce259f59_820d_41e4_8ef7_3b977664b7ea.slice/crio-230fa10a7317ce45644fb7fead959f7394a07f408ffac4c67cf4fc05835d91e5 WatchSource:0}: Error finding container 230fa10a7317ce45644fb7fead959f7394a07f408ffac4c67cf4fc05835d91e5: Status 404 returned error can't find the container with id 230fa10a7317ce45644fb7fead959f7394a07f408ffac4c67cf4fc05835d91e5 Oct 13 08:56:28 crc kubenswrapper[4685]: I1013 08:56:28.610584 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6f5fffbd86-pw7w6" event={"ID":"ce259f59-820d-41e4-8ef7-3b977664b7ea","Type":"ContainerStarted","Data":"230fa10a7317ce45644fb7fead959f7394a07f408ffac4c67cf4fc05835d91e5"} Oct 13 08:56:28 crc kubenswrapper[4685]: I1013 08:56:28.612334 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-759fc95d5d-79ckz" event={"ID":"b3d2fe67-48db-4dbc-869d-d0c13b18ec8a","Type":"ContainerStarted","Data":"bd5b6bca22073a4ad518262465916489fc343c2a2b8aecf62c8b9bd02b95bac5"} Oct 13 08:56:34 crc kubenswrapper[4685]: I1013 08:56:34.648104 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6f5fffbd86-pw7w6" event={"ID":"ce259f59-820d-41e4-8ef7-3b977664b7ea","Type":"ContainerStarted","Data":"4cbfeb2ad51bfec119ed905f908e0f2f5a65d55ac351b7290a448729982244e8"} Oct 13 08:56:34 crc kubenswrapper[4685]: I1013 08:56:34.648884 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6f5fffbd86-pw7w6" Oct 13 08:56:34 crc kubenswrapper[4685]: I1013 08:56:34.651198 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-759fc95d5d-79ckz" event={"ID":"b3d2fe67-48db-4dbc-869d-d0c13b18ec8a","Type":"ContainerStarted","Data":"a8f3441e806194ac3b0a12efa3a2240897f35a2d5af72dd5b087db6003b7bc23"} Oct 13 08:56:34 crc kubenswrapper[4685]: I1013 08:56:34.651336 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-759fc95d5d-79ckz" Oct 13 08:56:34 crc kubenswrapper[4685]: I1013 08:56:34.704713 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6f5fffbd86-pw7w6" podStartSLOduration=1.802941197 podStartE2EDuration="7.704697262s" podCreationTimestamp="2025-10-13 08:56:27 +0000 UTC" firstStartedPulling="2025-10-13 08:56:27.918598318 +0000 UTC m=+713.066474079" lastFinishedPulling="2025-10-13 08:56:33.820354373 +0000 UTC m=+718.968230144" observedRunningTime="2025-10-13 08:56:34.674117533 +0000 UTC m=+719.821993304" watchObservedRunningTime="2025-10-13 08:56:34.704697262 +0000 UTC m=+719.852573023" Oct 13 08:56:34 crc kubenswrapper[4685]: I1013 08:56:34.706580 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-759fc95d5d-79ckz" podStartSLOduration=2.500121773 podStartE2EDuration="8.706575302s" podCreationTimestamp="2025-10-13 08:56:26 +0000 UTC" firstStartedPulling="2025-10-13 08:56:27.607156059 +0000 UTC m=+712.755031820" lastFinishedPulling="2025-10-13 08:56:33.813609588 +0000 UTC m=+718.961485349" observedRunningTime="2025-10-13 08:56:34.698073229 +0000 UTC m=+719.845948990" watchObservedRunningTime="2025-10-13 08:56:34.706575302 +0000 UTC m=+719.854451063" Oct 13 08:56:47 crc kubenswrapper[4685]: I1013 08:56:47.674068 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6f5fffbd86-pw7w6" Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.273530 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9z47r"] Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.275346 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" podUID="7a81f8b4-7529-45f7-8ea5-df3252d70653" containerName="controller-manager" containerID="cri-o://a66adfc40f59a9f4657f087d72d56f04b48cd8dce962d4a760458c5c54315485" gracePeriod=30 Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.378191 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b"] Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.378388 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" podUID="b2158024-4bf6-457d-809d-c389b8ff2309" containerName="route-controller-manager" containerID="cri-o://53c78249cb47a22e9d6b54303c1e84c19d3830dcd507d0e79734d799ca546eca" gracePeriod=30 Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.788824 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.838078 4685 generic.go:334] "Generic (PLEG): container finished" podID="7a81f8b4-7529-45f7-8ea5-df3252d70653" containerID="a66adfc40f59a9f4657f087d72d56f04b48cd8dce962d4a760458c5c54315485" exitCode=0 Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.838171 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" event={"ID":"7a81f8b4-7529-45f7-8ea5-df3252d70653","Type":"ContainerDied","Data":"a66adfc40f59a9f4657f087d72d56f04b48cd8dce962d4a760458c5c54315485"} Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.838214 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" event={"ID":"7a81f8b4-7529-45f7-8ea5-df3252d70653","Type":"ContainerDied","Data":"20b415acc33a00de655d765eec169b0519ed69c6923a8371aac4eb82f8cf502a"} Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.838237 4685 scope.go:117] "RemoveContainer" containerID="a66adfc40f59a9f4657f087d72d56f04b48cd8dce962d4a760458c5c54315485" Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.838394 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9z47r" Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.841628 4685 generic.go:334] "Generic (PLEG): container finished" podID="b2158024-4bf6-457d-809d-c389b8ff2309" containerID="53c78249cb47a22e9d6b54303c1e84c19d3830dcd507d0e79734d799ca546eca" exitCode=0 Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.841700 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" event={"ID":"b2158024-4bf6-457d-809d-c389b8ff2309","Type":"ContainerDied","Data":"53c78249cb47a22e9d6b54303c1e84c19d3830dcd507d0e79734d799ca546eca"} Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.862985 4685 scope.go:117] "RemoveContainer" containerID="a66adfc40f59a9f4657f087d72d56f04b48cd8dce962d4a760458c5c54315485" Oct 13 08:57:01 crc kubenswrapper[4685]: E1013 08:57:01.863518 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a66adfc40f59a9f4657f087d72d56f04b48cd8dce962d4a760458c5c54315485\": container with ID starting with a66adfc40f59a9f4657f087d72d56f04b48cd8dce962d4a760458c5c54315485 not found: ID does not exist" containerID="a66adfc40f59a9f4657f087d72d56f04b48cd8dce962d4a760458c5c54315485" Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.863588 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a66adfc40f59a9f4657f087d72d56f04b48cd8dce962d4a760458c5c54315485"} err="failed to get container status \"a66adfc40f59a9f4657f087d72d56f04b48cd8dce962d4a760458c5c54315485\": rpc error: code = NotFound desc = could not find container \"a66adfc40f59a9f4657f087d72d56f04b48cd8dce962d4a760458c5c54315485\": container with ID starting with a66adfc40f59a9f4657f087d72d56f04b48cd8dce962d4a760458c5c54315485 not found: ID does not exist" Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.906069 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nh6nw\" (UniqueName: \"kubernetes.io/projected/7a81f8b4-7529-45f7-8ea5-df3252d70653-kube-api-access-nh6nw\") pod \"7a81f8b4-7529-45f7-8ea5-df3252d70653\" (UID: \"7a81f8b4-7529-45f7-8ea5-df3252d70653\") " Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.906157 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a81f8b4-7529-45f7-8ea5-df3252d70653-serving-cert\") pod \"7a81f8b4-7529-45f7-8ea5-df3252d70653\" (UID: \"7a81f8b4-7529-45f7-8ea5-df3252d70653\") " Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.906178 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7a81f8b4-7529-45f7-8ea5-df3252d70653-proxy-ca-bundles\") pod \"7a81f8b4-7529-45f7-8ea5-df3252d70653\" (UID: \"7a81f8b4-7529-45f7-8ea5-df3252d70653\") " Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.906214 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7a81f8b4-7529-45f7-8ea5-df3252d70653-client-ca\") pod \"7a81f8b4-7529-45f7-8ea5-df3252d70653\" (UID: \"7a81f8b4-7529-45f7-8ea5-df3252d70653\") " Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.906273 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a81f8b4-7529-45f7-8ea5-df3252d70653-config\") pod \"7a81f8b4-7529-45f7-8ea5-df3252d70653\" (UID: \"7a81f8b4-7529-45f7-8ea5-df3252d70653\") " Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.907376 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a81f8b4-7529-45f7-8ea5-df3252d70653-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7a81f8b4-7529-45f7-8ea5-df3252d70653" (UID: "7a81f8b4-7529-45f7-8ea5-df3252d70653"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.907394 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a81f8b4-7529-45f7-8ea5-df3252d70653-client-ca" (OuterVolumeSpecName: "client-ca") pod "7a81f8b4-7529-45f7-8ea5-df3252d70653" (UID: "7a81f8b4-7529-45f7-8ea5-df3252d70653"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.907577 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a81f8b4-7529-45f7-8ea5-df3252d70653-config" (OuterVolumeSpecName: "config") pod "7a81f8b4-7529-45f7-8ea5-df3252d70653" (UID: "7a81f8b4-7529-45f7-8ea5-df3252d70653"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.913762 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a81f8b4-7529-45f7-8ea5-df3252d70653-kube-api-access-nh6nw" (OuterVolumeSpecName: "kube-api-access-nh6nw") pod "7a81f8b4-7529-45f7-8ea5-df3252d70653" (UID: "7a81f8b4-7529-45f7-8ea5-df3252d70653"). InnerVolumeSpecName "kube-api-access-nh6nw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:57:01 crc kubenswrapper[4685]: I1013 08:57:01.915081 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a81f8b4-7529-45f7-8ea5-df3252d70653-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7a81f8b4-7529-45f7-8ea5-df3252d70653" (UID: "7a81f8b4-7529-45f7-8ea5-df3252d70653"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:57:02 crc kubenswrapper[4685]: I1013 08:57:02.007892 4685 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7a81f8b4-7529-45f7-8ea5-df3252d70653-client-ca\") on node \"crc\" DevicePath \"\"" Oct 13 08:57:02 crc kubenswrapper[4685]: I1013 08:57:02.007975 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a81f8b4-7529-45f7-8ea5-df3252d70653-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:57:02 crc kubenswrapper[4685]: I1013 08:57:02.007997 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nh6nw\" (UniqueName: \"kubernetes.io/projected/7a81f8b4-7529-45f7-8ea5-df3252d70653-kube-api-access-nh6nw\") on node \"crc\" DevicePath \"\"" Oct 13 08:57:02 crc kubenswrapper[4685]: I1013 08:57:02.008019 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a81f8b4-7529-45f7-8ea5-df3252d70653-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:57:02 crc kubenswrapper[4685]: I1013 08:57:02.008148 4685 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7a81f8b4-7529-45f7-8ea5-df3252d70653-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 13 08:57:02 crc kubenswrapper[4685]: I1013 08:57:02.169490 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9z47r"] Oct 13 08:57:02 crc kubenswrapper[4685]: I1013 08:57:02.173163 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9z47r"] Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.248470 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5dff79b858-2bkl7"] Oct 13 08:57:03 crc kubenswrapper[4685]: E1013 08:57:03.249126 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a81f8b4-7529-45f7-8ea5-df3252d70653" containerName="controller-manager" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.249145 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a81f8b4-7529-45f7-8ea5-df3252d70653" containerName="controller-manager" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.249324 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a81f8b4-7529-45f7-8ea5-df3252d70653" containerName="controller-manager" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.249864 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5dff79b858-2bkl7" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.258870 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.259283 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.262384 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.262690 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.262865 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.264036 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.264135 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5dff79b858-2bkl7"] Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.271686 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.326861 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6djn\" (UniqueName: \"kubernetes.io/projected/d316c6ee-087d-4c59-9453-0d6691516179-kube-api-access-x6djn\") pod \"controller-manager-5dff79b858-2bkl7\" (UID: \"d316c6ee-087d-4c59-9453-0d6691516179\") " pod="openshift-controller-manager/controller-manager-5dff79b858-2bkl7" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.327117 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d316c6ee-087d-4c59-9453-0d6691516179-serving-cert\") pod \"controller-manager-5dff79b858-2bkl7\" (UID: \"d316c6ee-087d-4c59-9453-0d6691516179\") " pod="openshift-controller-manager/controller-manager-5dff79b858-2bkl7" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.327322 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d316c6ee-087d-4c59-9453-0d6691516179-config\") pod \"controller-manager-5dff79b858-2bkl7\" (UID: \"d316c6ee-087d-4c59-9453-0d6691516179\") " pod="openshift-controller-manager/controller-manager-5dff79b858-2bkl7" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.327590 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d316c6ee-087d-4c59-9453-0d6691516179-proxy-ca-bundles\") pod \"controller-manager-5dff79b858-2bkl7\" (UID: \"d316c6ee-087d-4c59-9453-0d6691516179\") " pod="openshift-controller-manager/controller-manager-5dff79b858-2bkl7" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.327709 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d316c6ee-087d-4c59-9453-0d6691516179-client-ca\") pod \"controller-manager-5dff79b858-2bkl7\" (UID: \"d316c6ee-087d-4c59-9453-0d6691516179\") " pod="openshift-controller-manager/controller-manager-5dff79b858-2bkl7" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.429499 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d316c6ee-087d-4c59-9453-0d6691516179-serving-cert\") pod \"controller-manager-5dff79b858-2bkl7\" (UID: \"d316c6ee-087d-4c59-9453-0d6691516179\") " pod="openshift-controller-manager/controller-manager-5dff79b858-2bkl7" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.429574 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d316c6ee-087d-4c59-9453-0d6691516179-config\") pod \"controller-manager-5dff79b858-2bkl7\" (UID: \"d316c6ee-087d-4c59-9453-0d6691516179\") " pod="openshift-controller-manager/controller-manager-5dff79b858-2bkl7" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.429634 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d316c6ee-087d-4c59-9453-0d6691516179-proxy-ca-bundles\") pod \"controller-manager-5dff79b858-2bkl7\" (UID: \"d316c6ee-087d-4c59-9453-0d6691516179\") " pod="openshift-controller-manager/controller-manager-5dff79b858-2bkl7" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.429672 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d316c6ee-087d-4c59-9453-0d6691516179-client-ca\") pod \"controller-manager-5dff79b858-2bkl7\" (UID: \"d316c6ee-087d-4c59-9453-0d6691516179\") " pod="openshift-controller-manager/controller-manager-5dff79b858-2bkl7" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.429707 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6djn\" (UniqueName: \"kubernetes.io/projected/d316c6ee-087d-4c59-9453-0d6691516179-kube-api-access-x6djn\") pod \"controller-manager-5dff79b858-2bkl7\" (UID: \"d316c6ee-087d-4c59-9453-0d6691516179\") " pod="openshift-controller-manager/controller-manager-5dff79b858-2bkl7" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.431081 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d316c6ee-087d-4c59-9453-0d6691516179-client-ca\") pod \"controller-manager-5dff79b858-2bkl7\" (UID: \"d316c6ee-087d-4c59-9453-0d6691516179\") " pod="openshift-controller-manager/controller-manager-5dff79b858-2bkl7" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.431308 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d316c6ee-087d-4c59-9453-0d6691516179-config\") pod \"controller-manager-5dff79b858-2bkl7\" (UID: \"d316c6ee-087d-4c59-9453-0d6691516179\") " pod="openshift-controller-manager/controller-manager-5dff79b858-2bkl7" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.431861 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d316c6ee-087d-4c59-9453-0d6691516179-proxy-ca-bundles\") pod \"controller-manager-5dff79b858-2bkl7\" (UID: \"d316c6ee-087d-4c59-9453-0d6691516179\") " pod="openshift-controller-manager/controller-manager-5dff79b858-2bkl7" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.436307 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d316c6ee-087d-4c59-9453-0d6691516179-serving-cert\") pod \"controller-manager-5dff79b858-2bkl7\" (UID: \"d316c6ee-087d-4c59-9453-0d6691516179\") " pod="openshift-controller-manager/controller-manager-5dff79b858-2bkl7" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.458278 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6djn\" (UniqueName: \"kubernetes.io/projected/d316c6ee-087d-4c59-9453-0d6691516179-kube-api-access-x6djn\") pod \"controller-manager-5dff79b858-2bkl7\" (UID: \"d316c6ee-087d-4c59-9453-0d6691516179\") " pod="openshift-controller-manager/controller-manager-5dff79b858-2bkl7" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.510378 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a81f8b4-7529-45f7-8ea5-df3252d70653" path="/var/lib/kubelet/pods/7a81f8b4-7529-45f7-8ea5-df3252d70653/volumes" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.611490 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5dff79b858-2bkl7" Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.828824 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5dff79b858-2bkl7"] Oct 13 08:57:03 crc kubenswrapper[4685]: I1013 08:57:03.864463 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5dff79b858-2bkl7" event={"ID":"d316c6ee-087d-4c59-9453-0d6691516179","Type":"ContainerStarted","Data":"cf01ec300d663a45baa8f554737da0d3ed3bd813fc324357891ba9dc80e39859"} Oct 13 08:57:05 crc kubenswrapper[4685]: I1013 08:57:05.466202 4685 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-ssk8b container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Oct 13 08:57:05 crc kubenswrapper[4685]: I1013 08:57:05.466494 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" podUID="b2158024-4bf6-457d-809d-c389b8ff2309" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.286557 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-759fc95d5d-79ckz" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.642983 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.695587 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x"] Oct 13 08:57:07 crc kubenswrapper[4685]: E1013 08:57:07.696189 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2158024-4bf6-457d-809d-c389b8ff2309" containerName="route-controller-manager" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.696210 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2158024-4bf6-457d-809d-c389b8ff2309" containerName="route-controller-manager" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.696227 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b2158024-4bf6-457d-809d-c389b8ff2309-client-ca\") pod \"b2158024-4bf6-457d-809d-c389b8ff2309\" (UID: \"b2158024-4bf6-457d-809d-c389b8ff2309\") " Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.696280 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2158024-4bf6-457d-809d-c389b8ff2309-config\") pod \"b2158024-4bf6-457d-809d-c389b8ff2309\" (UID: \"b2158024-4bf6-457d-809d-c389b8ff2309\") " Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.696328 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2158024-4bf6-457d-809d-c389b8ff2309-serving-cert\") pod \"b2158024-4bf6-457d-809d-c389b8ff2309\" (UID: \"b2158024-4bf6-457d-809d-c389b8ff2309\") " Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.696408 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpqnp\" (UniqueName: \"kubernetes.io/projected/b2158024-4bf6-457d-809d-c389b8ff2309-kube-api-access-mpqnp\") pod \"b2158024-4bf6-457d-809d-c389b8ff2309\" (UID: \"b2158024-4bf6-457d-809d-c389b8ff2309\") " Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.696529 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2158024-4bf6-457d-809d-c389b8ff2309" containerName="route-controller-manager" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.697062 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2158024-4bf6-457d-809d-c389b8ff2309-client-ca" (OuterVolumeSpecName: "client-ca") pod "b2158024-4bf6-457d-809d-c389b8ff2309" (UID: "b2158024-4bf6-457d-809d-c389b8ff2309"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.697112 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2158024-4bf6-457d-809d-c389b8ff2309-config" (OuterVolumeSpecName: "config") pod "b2158024-4bf6-457d-809d-c389b8ff2309" (UID: "b2158024-4bf6-457d-809d-c389b8ff2309"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.697253 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.702574 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x"] Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.708635 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2158024-4bf6-457d-809d-c389b8ff2309-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b2158024-4bf6-457d-809d-c389b8ff2309" (UID: "b2158024-4bf6-457d-809d-c389b8ff2309"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.723943 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2158024-4bf6-457d-809d-c389b8ff2309-kube-api-access-mpqnp" (OuterVolumeSpecName: "kube-api-access-mpqnp") pod "b2158024-4bf6-457d-809d-c389b8ff2309" (UID: "b2158024-4bf6-457d-809d-c389b8ff2309"). InnerVolumeSpecName "kube-api-access-mpqnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.798226 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tn6sd\" (UniqueName: \"kubernetes.io/projected/de0fea9f-9a07-43c6-91af-61f8ef5957d7-kube-api-access-tn6sd\") pod \"route-controller-manager-644446c6b6-pb77x\" (UID: \"de0fea9f-9a07-43c6-91af-61f8ef5957d7\") " pod="openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.798282 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/de0fea9f-9a07-43c6-91af-61f8ef5957d7-client-ca\") pod \"route-controller-manager-644446c6b6-pb77x\" (UID: \"de0fea9f-9a07-43c6-91af-61f8ef5957d7\") " pod="openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.798337 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de0fea9f-9a07-43c6-91af-61f8ef5957d7-serving-cert\") pod \"route-controller-manager-644446c6b6-pb77x\" (UID: \"de0fea9f-9a07-43c6-91af-61f8ef5957d7\") " pod="openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.798373 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de0fea9f-9a07-43c6-91af-61f8ef5957d7-config\") pod \"route-controller-manager-644446c6b6-pb77x\" (UID: \"de0fea9f-9a07-43c6-91af-61f8ef5957d7\") " pod="openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.798423 4685 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b2158024-4bf6-457d-809d-c389b8ff2309-client-ca\") on node \"crc\" DevicePath \"\"" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.798438 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2158024-4bf6-457d-809d-c389b8ff2309-config\") on node \"crc\" DevicePath \"\"" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.798449 4685 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2158024-4bf6-457d-809d-c389b8ff2309-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.798459 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpqnp\" (UniqueName: \"kubernetes.io/projected/b2158024-4bf6-457d-809d-c389b8ff2309-kube-api-access-mpqnp\") on node \"crc\" DevicePath \"\"" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.885315 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" event={"ID":"b2158024-4bf6-457d-809d-c389b8ff2309","Type":"ContainerDied","Data":"612cca0a022e01375aed7ba5cf010b2ccef81567dd9db1a92f3e0ca7ed53c94a"} Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.885366 4685 scope.go:117] "RemoveContainer" containerID="53c78249cb47a22e9d6b54303c1e84c19d3830dcd507d0e79734d799ca546eca" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.885667 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.889348 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5dff79b858-2bkl7" event={"ID":"d316c6ee-087d-4c59-9453-0d6691516179","Type":"ContainerStarted","Data":"8344fb987d045d6eb0f64ab785f634835c8c35f9c5935113c373707a9c222cf9"} Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.891300 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5dff79b858-2bkl7" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.896485 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5dff79b858-2bkl7" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.899005 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de0fea9f-9a07-43c6-91af-61f8ef5957d7-serving-cert\") pod \"route-controller-manager-644446c6b6-pb77x\" (UID: \"de0fea9f-9a07-43c6-91af-61f8ef5957d7\") " pod="openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.899044 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de0fea9f-9a07-43c6-91af-61f8ef5957d7-config\") pod \"route-controller-manager-644446c6b6-pb77x\" (UID: \"de0fea9f-9a07-43c6-91af-61f8ef5957d7\") " pod="openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.899089 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tn6sd\" (UniqueName: \"kubernetes.io/projected/de0fea9f-9a07-43c6-91af-61f8ef5957d7-kube-api-access-tn6sd\") pod \"route-controller-manager-644446c6b6-pb77x\" (UID: \"de0fea9f-9a07-43c6-91af-61f8ef5957d7\") " pod="openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.899111 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/de0fea9f-9a07-43c6-91af-61f8ef5957d7-client-ca\") pod \"route-controller-manager-644446c6b6-pb77x\" (UID: \"de0fea9f-9a07-43c6-91af-61f8ef5957d7\") " pod="openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.899905 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/de0fea9f-9a07-43c6-91af-61f8ef5957d7-client-ca\") pod \"route-controller-manager-644446c6b6-pb77x\" (UID: \"de0fea9f-9a07-43c6-91af-61f8ef5957d7\") " pod="openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.902345 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de0fea9f-9a07-43c6-91af-61f8ef5957d7-config\") pod \"route-controller-manager-644446c6b6-pb77x\" (UID: \"de0fea9f-9a07-43c6-91af-61f8ef5957d7\") " pod="openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.915775 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de0fea9f-9a07-43c6-91af-61f8ef5957d7-serving-cert\") pod \"route-controller-manager-644446c6b6-pb77x\" (UID: \"de0fea9f-9a07-43c6-91af-61f8ef5957d7\") " pod="openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.929372 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5dff79b858-2bkl7" podStartSLOduration=6.929350915 podStartE2EDuration="6.929350915s" podCreationTimestamp="2025-10-13 08:57:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:57:07.927706709 +0000 UTC m=+753.075582470" watchObservedRunningTime="2025-10-13 08:57:07.929350915 +0000 UTC m=+753.077226676" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.953963 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b"] Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.954606 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tn6sd\" (UniqueName: \"kubernetes.io/projected/de0fea9f-9a07-43c6-91af-61f8ef5957d7-kube-api-access-tn6sd\") pod \"route-controller-manager-644446c6b6-pb77x\" (UID: \"de0fea9f-9a07-43c6-91af-61f8ef5957d7\") " pod="openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x" Oct 13 08:57:07 crc kubenswrapper[4685]: I1013 08:57:07.969842 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ssk8b"] Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.055150 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.095487 4685 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.320416 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-ggf8f"] Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.328494 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.340717 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-fkd9z"] Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.343554 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.347317 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-gdm47" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.348670 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-fkd9z" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.348893 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.362812 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.364873 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-fkd9z"] Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.406491 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/33a86974-4a79-42b8-8bf9-cf3992b9b54d-frr-conf\") pod \"frr-k8s-ggf8f\" (UID: \"33a86974-4a79-42b8-8bf9-cf3992b9b54d\") " pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.406535 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/33a86974-4a79-42b8-8bf9-cf3992b9b54d-metrics-certs\") pod \"frr-k8s-ggf8f\" (UID: \"33a86974-4a79-42b8-8bf9-cf3992b9b54d\") " pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.406555 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfz2p\" (UniqueName: \"kubernetes.io/projected/9b097f6c-ec81-4342-b4c1-63520267ba08-kube-api-access-nfz2p\") pod \"frr-k8s-webhook-server-64bf5d555-fkd9z\" (UID: \"9b097f6c-ec81-4342-b4c1-63520267ba08\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-fkd9z" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.406652 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/33a86974-4a79-42b8-8bf9-cf3992b9b54d-frr-sockets\") pod \"frr-k8s-ggf8f\" (UID: \"33a86974-4a79-42b8-8bf9-cf3992b9b54d\") " pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.406694 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9b097f6c-ec81-4342-b4c1-63520267ba08-cert\") pod \"frr-k8s-webhook-server-64bf5d555-fkd9z\" (UID: \"9b097f6c-ec81-4342-b4c1-63520267ba08\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-fkd9z" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.406715 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d75zm\" (UniqueName: \"kubernetes.io/projected/33a86974-4a79-42b8-8bf9-cf3992b9b54d-kube-api-access-d75zm\") pod \"frr-k8s-ggf8f\" (UID: \"33a86974-4a79-42b8-8bf9-cf3992b9b54d\") " pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.406737 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/33a86974-4a79-42b8-8bf9-cf3992b9b54d-metrics\") pod \"frr-k8s-ggf8f\" (UID: \"33a86974-4a79-42b8-8bf9-cf3992b9b54d\") " pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.406764 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/33a86974-4a79-42b8-8bf9-cf3992b9b54d-frr-startup\") pod \"frr-k8s-ggf8f\" (UID: \"33a86974-4a79-42b8-8bf9-cf3992b9b54d\") " pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.406797 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/33a86974-4a79-42b8-8bf9-cf3992b9b54d-reloader\") pod \"frr-k8s-ggf8f\" (UID: \"33a86974-4a79-42b8-8bf9-cf3992b9b54d\") " pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.503386 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-ddb2s"] Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.504208 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-ddb2s" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.511571 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.511761 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.511961 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-787ch" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.512084 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.512787 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/33a86974-4a79-42b8-8bf9-cf3992b9b54d-frr-conf\") pod \"frr-k8s-ggf8f\" (UID: \"33a86974-4a79-42b8-8bf9-cf3992b9b54d\") " pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.512817 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/33a86974-4a79-42b8-8bf9-cf3992b9b54d-metrics-certs\") pod \"frr-k8s-ggf8f\" (UID: \"33a86974-4a79-42b8-8bf9-cf3992b9b54d\") " pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.512834 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfz2p\" (UniqueName: \"kubernetes.io/projected/9b097f6c-ec81-4342-b4c1-63520267ba08-kube-api-access-nfz2p\") pod \"frr-k8s-webhook-server-64bf5d555-fkd9z\" (UID: \"9b097f6c-ec81-4342-b4c1-63520267ba08\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-fkd9z" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.512883 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/33a86974-4a79-42b8-8bf9-cf3992b9b54d-frr-sockets\") pod \"frr-k8s-ggf8f\" (UID: \"33a86974-4a79-42b8-8bf9-cf3992b9b54d\") " pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.512901 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9b097f6c-ec81-4342-b4c1-63520267ba08-cert\") pod \"frr-k8s-webhook-server-64bf5d555-fkd9z\" (UID: \"9b097f6c-ec81-4342-b4c1-63520267ba08\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-fkd9z" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.512950 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d75zm\" (UniqueName: \"kubernetes.io/projected/33a86974-4a79-42b8-8bf9-cf3992b9b54d-kube-api-access-d75zm\") pod \"frr-k8s-ggf8f\" (UID: \"33a86974-4a79-42b8-8bf9-cf3992b9b54d\") " pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.512968 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/33a86974-4a79-42b8-8bf9-cf3992b9b54d-metrics\") pod \"frr-k8s-ggf8f\" (UID: \"33a86974-4a79-42b8-8bf9-cf3992b9b54d\") " pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.512987 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/33a86974-4a79-42b8-8bf9-cf3992b9b54d-frr-startup\") pod \"frr-k8s-ggf8f\" (UID: \"33a86974-4a79-42b8-8bf9-cf3992b9b54d\") " pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.513007 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/33a86974-4a79-42b8-8bf9-cf3992b9b54d-reloader\") pod \"frr-k8s-ggf8f\" (UID: \"33a86974-4a79-42b8-8bf9-cf3992b9b54d\") " pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: E1013 08:57:08.513678 4685 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 13 08:57:08 crc kubenswrapper[4685]: E1013 08:57:08.513748 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9b097f6c-ec81-4342-b4c1-63520267ba08-cert podName:9b097f6c-ec81-4342-b4c1-63520267ba08 nodeName:}" failed. No retries permitted until 2025-10-13 08:57:09.013730578 +0000 UTC m=+754.161606329 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9b097f6c-ec81-4342-b4c1-63520267ba08-cert") pod "frr-k8s-webhook-server-64bf5d555-fkd9z" (UID: "9b097f6c-ec81-4342-b4c1-63520267ba08") : secret "frr-k8s-webhook-server-cert" not found Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.515248 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/33a86974-4a79-42b8-8bf9-cf3992b9b54d-reloader\") pod \"frr-k8s-ggf8f\" (UID: \"33a86974-4a79-42b8-8bf9-cf3992b9b54d\") " pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.518321 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/33a86974-4a79-42b8-8bf9-cf3992b9b54d-metrics-certs\") pod \"frr-k8s-ggf8f\" (UID: \"33a86974-4a79-42b8-8bf9-cf3992b9b54d\") " pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.519301 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/33a86974-4a79-42b8-8bf9-cf3992b9b54d-frr-startup\") pod \"frr-k8s-ggf8f\" (UID: \"33a86974-4a79-42b8-8bf9-cf3992b9b54d\") " pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.519520 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/33a86974-4a79-42b8-8bf9-cf3992b9b54d-frr-conf\") pod \"frr-k8s-ggf8f\" (UID: \"33a86974-4a79-42b8-8bf9-cf3992b9b54d\") " pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.519984 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/33a86974-4a79-42b8-8bf9-cf3992b9b54d-metrics\") pod \"frr-k8s-ggf8f\" (UID: \"33a86974-4a79-42b8-8bf9-cf3992b9b54d\") " pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.532613 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/33a86974-4a79-42b8-8bf9-cf3992b9b54d-frr-sockets\") pod \"frr-k8s-ggf8f\" (UID: \"33a86974-4a79-42b8-8bf9-cf3992b9b54d\") " pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.535467 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfz2p\" (UniqueName: \"kubernetes.io/projected/9b097f6c-ec81-4342-b4c1-63520267ba08-kube-api-access-nfz2p\") pod \"frr-k8s-webhook-server-64bf5d555-fkd9z\" (UID: \"9b097f6c-ec81-4342-b4c1-63520267ba08\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-fkd9z" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.536042 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-fmsz9"] Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.536900 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-fmsz9" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.540612 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.550226 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-fmsz9"] Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.572072 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d75zm\" (UniqueName: \"kubernetes.io/projected/33a86974-4a79-42b8-8bf9-cf3992b9b54d-kube-api-access-d75zm\") pod \"frr-k8s-ggf8f\" (UID: \"33a86974-4a79-42b8-8bf9-cf3992b9b54d\") " pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.616577 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/55ccd123-f671-4230-b2d5-e6ffb265429a-memberlist\") pod \"speaker-ddb2s\" (UID: \"55ccd123-f671-4230-b2d5-e6ffb265429a\") " pod="metallb-system/speaker-ddb2s" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.616749 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/55ccd123-f671-4230-b2d5-e6ffb265429a-metallb-excludel2\") pod \"speaker-ddb2s\" (UID: \"55ccd123-f671-4230-b2d5-e6ffb265429a\") " pod="metallb-system/speaker-ddb2s" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.616845 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55ccd123-f671-4230-b2d5-e6ffb265429a-metrics-certs\") pod \"speaker-ddb2s\" (UID: \"55ccd123-f671-4230-b2d5-e6ffb265429a\") " pod="metallb-system/speaker-ddb2s" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.616934 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szrtv\" (UniqueName: \"kubernetes.io/projected/55ccd123-f671-4230-b2d5-e6ffb265429a-kube-api-access-szrtv\") pod \"speaker-ddb2s\" (UID: \"55ccd123-f671-4230-b2d5-e6ffb265429a\") " pod="metallb-system/speaker-ddb2s" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.617038 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34bb5061-30c8-49f5-8dca-a411b3f128b2-metrics-certs\") pod \"controller-68d546b9d8-fmsz9\" (UID: \"34bb5061-30c8-49f5-8dca-a411b3f128b2\") " pod="metallb-system/controller-68d546b9d8-fmsz9" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.617114 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4ck9\" (UniqueName: \"kubernetes.io/projected/34bb5061-30c8-49f5-8dca-a411b3f128b2-kube-api-access-w4ck9\") pod \"controller-68d546b9d8-fmsz9\" (UID: \"34bb5061-30c8-49f5-8dca-a411b3f128b2\") " pod="metallb-system/controller-68d546b9d8-fmsz9" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.617179 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/34bb5061-30c8-49f5-8dca-a411b3f128b2-cert\") pod \"controller-68d546b9d8-fmsz9\" (UID: \"34bb5061-30c8-49f5-8dca-a411b3f128b2\") " pod="metallb-system/controller-68d546b9d8-fmsz9" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.677375 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.718170 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/34bb5061-30c8-49f5-8dca-a411b3f128b2-cert\") pod \"controller-68d546b9d8-fmsz9\" (UID: \"34bb5061-30c8-49f5-8dca-a411b3f128b2\") " pod="metallb-system/controller-68d546b9d8-fmsz9" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.718567 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/55ccd123-f671-4230-b2d5-e6ffb265429a-memberlist\") pod \"speaker-ddb2s\" (UID: \"55ccd123-f671-4230-b2d5-e6ffb265429a\") " pod="metallb-system/speaker-ddb2s" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.718589 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/55ccd123-f671-4230-b2d5-e6ffb265429a-metallb-excludel2\") pod \"speaker-ddb2s\" (UID: \"55ccd123-f671-4230-b2d5-e6ffb265429a\") " pod="metallb-system/speaker-ddb2s" Oct 13 08:57:08 crc kubenswrapper[4685]: E1013 08:57:08.718657 4685 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 13 08:57:08 crc kubenswrapper[4685]: E1013 08:57:08.718698 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55ccd123-f671-4230-b2d5-e6ffb265429a-memberlist podName:55ccd123-f671-4230-b2d5-e6ffb265429a nodeName:}" failed. No retries permitted until 2025-10-13 08:57:09.218685748 +0000 UTC m=+754.366561499 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/55ccd123-f671-4230-b2d5-e6ffb265429a-memberlist") pod "speaker-ddb2s" (UID: "55ccd123-f671-4230-b2d5-e6ffb265429a") : secret "metallb-memberlist" not found Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.718754 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55ccd123-f671-4230-b2d5-e6ffb265429a-metrics-certs\") pod \"speaker-ddb2s\" (UID: \"55ccd123-f671-4230-b2d5-e6ffb265429a\") " pod="metallb-system/speaker-ddb2s" Oct 13 08:57:08 crc kubenswrapper[4685]: E1013 08:57:08.719223 4685 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Oct 13 08:57:08 crc kubenswrapper[4685]: E1013 08:57:08.719330 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55ccd123-f671-4230-b2d5-e6ffb265429a-metrics-certs podName:55ccd123-f671-4230-b2d5-e6ffb265429a nodeName:}" failed. No retries permitted until 2025-10-13 08:57:09.219313555 +0000 UTC m=+754.367189316 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/55ccd123-f671-4230-b2d5-e6ffb265429a-metrics-certs") pod "speaker-ddb2s" (UID: "55ccd123-f671-4230-b2d5-e6ffb265429a") : secret "speaker-certs-secret" not found Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.719411 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/55ccd123-f671-4230-b2d5-e6ffb265429a-metallb-excludel2\") pod \"speaker-ddb2s\" (UID: \"55ccd123-f671-4230-b2d5-e6ffb265429a\") " pod="metallb-system/speaker-ddb2s" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.719480 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szrtv\" (UniqueName: \"kubernetes.io/projected/55ccd123-f671-4230-b2d5-e6ffb265429a-kube-api-access-szrtv\") pod \"speaker-ddb2s\" (UID: \"55ccd123-f671-4230-b2d5-e6ffb265429a\") " pod="metallb-system/speaker-ddb2s" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.719697 4685 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.719503 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34bb5061-30c8-49f5-8dca-a411b3f128b2-metrics-certs\") pod \"controller-68d546b9d8-fmsz9\" (UID: \"34bb5061-30c8-49f5-8dca-a411b3f128b2\") " pod="metallb-system/controller-68d546b9d8-fmsz9" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.720310 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4ck9\" (UniqueName: \"kubernetes.io/projected/34bb5061-30c8-49f5-8dca-a411b3f128b2-kube-api-access-w4ck9\") pod \"controller-68d546b9d8-fmsz9\" (UID: \"34bb5061-30c8-49f5-8dca-a411b3f128b2\") " pod="metallb-system/controller-68d546b9d8-fmsz9" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.730124 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34bb5061-30c8-49f5-8dca-a411b3f128b2-metrics-certs\") pod \"controller-68d546b9d8-fmsz9\" (UID: \"34bb5061-30c8-49f5-8dca-a411b3f128b2\") " pod="metallb-system/controller-68d546b9d8-fmsz9" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.736353 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/34bb5061-30c8-49f5-8dca-a411b3f128b2-cert\") pod \"controller-68d546b9d8-fmsz9\" (UID: \"34bb5061-30c8-49f5-8dca-a411b3f128b2\") " pod="metallb-system/controller-68d546b9d8-fmsz9" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.742354 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szrtv\" (UniqueName: \"kubernetes.io/projected/55ccd123-f671-4230-b2d5-e6ffb265429a-kube-api-access-szrtv\") pod \"speaker-ddb2s\" (UID: \"55ccd123-f671-4230-b2d5-e6ffb265429a\") " pod="metallb-system/speaker-ddb2s" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.742976 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4ck9\" (UniqueName: \"kubernetes.io/projected/34bb5061-30c8-49f5-8dca-a411b3f128b2-kube-api-access-w4ck9\") pod \"controller-68d546b9d8-fmsz9\" (UID: \"34bb5061-30c8-49f5-8dca-a411b3f128b2\") " pod="metallb-system/controller-68d546b9d8-fmsz9" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.750056 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x"] Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.857170 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-fmsz9" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.906116 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ggf8f" event={"ID":"33a86974-4a79-42b8-8bf9-cf3992b9b54d","Type":"ContainerStarted","Data":"6d37d415bbc4235e10f0e1e7e14084b02a8fd29b2ddbbee6633b3f1ac7e22f3d"} Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.908341 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x" event={"ID":"de0fea9f-9a07-43c6-91af-61f8ef5957d7","Type":"ContainerStarted","Data":"0895bcc5b3fc76d99cfb7efb4a865bca747fbaad90a9ec9bebe555a937ed7e74"} Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.908367 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x" event={"ID":"de0fea9f-9a07-43c6-91af-61f8ef5957d7","Type":"ContainerStarted","Data":"018f841d73550f3d5d19091b0a60f4323499a858ee9a9b735a0ad87d50ff4d9e"} Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.908602 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.909506 4685 patch_prober.go:28] interesting pod/route-controller-manager-644446c6b6-pb77x container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.50:8443/healthz\": dial tcp 10.217.0.50:8443: connect: connection refused" start-of-body= Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.909537 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x" podUID="de0fea9f-9a07-43c6-91af-61f8ef5957d7" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.50:8443/healthz\": dial tcp 10.217.0.50:8443: connect: connection refused" Oct 13 08:57:08 crc kubenswrapper[4685]: I1013 08:57:08.928523 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x" podStartSLOduration=7.92850441 podStartE2EDuration="7.92850441s" podCreationTimestamp="2025-10-13 08:57:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:57:08.926619359 +0000 UTC m=+754.074495120" watchObservedRunningTime="2025-10-13 08:57:08.92850441 +0000 UTC m=+754.076380161" Oct 13 08:57:09 crc kubenswrapper[4685]: I1013 08:57:09.023312 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9b097f6c-ec81-4342-b4c1-63520267ba08-cert\") pod \"frr-k8s-webhook-server-64bf5d555-fkd9z\" (UID: \"9b097f6c-ec81-4342-b4c1-63520267ba08\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-fkd9z" Oct 13 08:57:09 crc kubenswrapper[4685]: I1013 08:57:09.028699 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9b097f6c-ec81-4342-b4c1-63520267ba08-cert\") pod \"frr-k8s-webhook-server-64bf5d555-fkd9z\" (UID: \"9b097f6c-ec81-4342-b4c1-63520267ba08\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-fkd9z" Oct 13 08:57:09 crc kubenswrapper[4685]: I1013 08:57:09.225928 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/55ccd123-f671-4230-b2d5-e6ffb265429a-memberlist\") pod \"speaker-ddb2s\" (UID: \"55ccd123-f671-4230-b2d5-e6ffb265429a\") " pod="metallb-system/speaker-ddb2s" Oct 13 08:57:09 crc kubenswrapper[4685]: I1013 08:57:09.225977 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55ccd123-f671-4230-b2d5-e6ffb265429a-metrics-certs\") pod \"speaker-ddb2s\" (UID: \"55ccd123-f671-4230-b2d5-e6ffb265429a\") " pod="metallb-system/speaker-ddb2s" Oct 13 08:57:09 crc kubenswrapper[4685]: E1013 08:57:09.226026 4685 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 13 08:57:09 crc kubenswrapper[4685]: E1013 08:57:09.226079 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55ccd123-f671-4230-b2d5-e6ffb265429a-memberlist podName:55ccd123-f671-4230-b2d5-e6ffb265429a nodeName:}" failed. No retries permitted until 2025-10-13 08:57:10.22606454 +0000 UTC m=+755.373940301 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/55ccd123-f671-4230-b2d5-e6ffb265429a-memberlist") pod "speaker-ddb2s" (UID: "55ccd123-f671-4230-b2d5-e6ffb265429a") : secret "metallb-memberlist" not found Oct 13 08:57:09 crc kubenswrapper[4685]: I1013 08:57:09.230074 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55ccd123-f671-4230-b2d5-e6ffb265429a-metrics-certs\") pod \"speaker-ddb2s\" (UID: \"55ccd123-f671-4230-b2d5-e6ffb265429a\") " pod="metallb-system/speaker-ddb2s" Oct 13 08:57:09 crc kubenswrapper[4685]: I1013 08:57:09.286883 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-fkd9z" Oct 13 08:57:09 crc kubenswrapper[4685]: I1013 08:57:09.342686 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-fmsz9"] Oct 13 08:57:09 crc kubenswrapper[4685]: W1013 08:57:09.372327 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34bb5061_30c8_49f5_8dca_a411b3f128b2.slice/crio-259fe3372a8a87236372c0537a125b0f334cf31fb3c87a455a6dbffac5db8e54 WatchSource:0}: Error finding container 259fe3372a8a87236372c0537a125b0f334cf31fb3c87a455a6dbffac5db8e54: Status 404 returned error can't find the container with id 259fe3372a8a87236372c0537a125b0f334cf31fb3c87a455a6dbffac5db8e54 Oct 13 08:57:09 crc kubenswrapper[4685]: I1013 08:57:09.518120 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2158024-4bf6-457d-809d-c389b8ff2309" path="/var/lib/kubelet/pods/b2158024-4bf6-457d-809d-c389b8ff2309/volumes" Oct 13 08:57:09 crc kubenswrapper[4685]: I1013 08:57:09.679271 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-fkd9z"] Oct 13 08:57:09 crc kubenswrapper[4685]: I1013 08:57:09.914153 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-fmsz9" event={"ID":"34bb5061-30c8-49f5-8dca-a411b3f128b2","Type":"ContainerStarted","Data":"7c6806be4accb077b724ef558938effce08dd78bd5d7fb1a2bf9f5ac185f46b5"} Oct 13 08:57:09 crc kubenswrapper[4685]: I1013 08:57:09.914471 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-fmsz9" event={"ID":"34bb5061-30c8-49f5-8dca-a411b3f128b2","Type":"ContainerStarted","Data":"e59430221e27f9f059aa4198dd10a6b6142e855373b0ce914fd8293a914bdf28"} Oct 13 08:57:09 crc kubenswrapper[4685]: I1013 08:57:09.914483 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-fmsz9" event={"ID":"34bb5061-30c8-49f5-8dca-a411b3f128b2","Type":"ContainerStarted","Data":"259fe3372a8a87236372c0537a125b0f334cf31fb3c87a455a6dbffac5db8e54"} Oct 13 08:57:09 crc kubenswrapper[4685]: I1013 08:57:09.914721 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-fmsz9" Oct 13 08:57:09 crc kubenswrapper[4685]: I1013 08:57:09.916036 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-fkd9z" event={"ID":"9b097f6c-ec81-4342-b4c1-63520267ba08","Type":"ContainerStarted","Data":"e4f20844475146f859c436a1a1ea4daf7d4b06bec497eefde6bdb15f17bbd4f9"} Oct 13 08:57:09 crc kubenswrapper[4685]: I1013 08:57:09.921137 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-644446c6b6-pb77x" Oct 13 08:57:09 crc kubenswrapper[4685]: I1013 08:57:09.932510 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-fmsz9" podStartSLOduration=1.9324945 podStartE2EDuration="1.9324945s" podCreationTimestamp="2025-10-13 08:57:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:57:09.928574652 +0000 UTC m=+755.076450413" watchObservedRunningTime="2025-10-13 08:57:09.9324945 +0000 UTC m=+755.080370261" Oct 13 08:57:10 crc kubenswrapper[4685]: I1013 08:57:10.237562 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/55ccd123-f671-4230-b2d5-e6ffb265429a-memberlist\") pod \"speaker-ddb2s\" (UID: \"55ccd123-f671-4230-b2d5-e6ffb265429a\") " pod="metallb-system/speaker-ddb2s" Oct 13 08:57:10 crc kubenswrapper[4685]: I1013 08:57:10.243440 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/55ccd123-f671-4230-b2d5-e6ffb265429a-memberlist\") pod \"speaker-ddb2s\" (UID: \"55ccd123-f671-4230-b2d5-e6ffb265429a\") " pod="metallb-system/speaker-ddb2s" Oct 13 08:57:10 crc kubenswrapper[4685]: I1013 08:57:10.336800 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-ddb2s" Oct 13 08:57:10 crc kubenswrapper[4685]: I1013 08:57:10.924814 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-ddb2s" event={"ID":"55ccd123-f671-4230-b2d5-e6ffb265429a","Type":"ContainerStarted","Data":"85abf1efff95edada5bf970afe59e69ccf616c5b356c27f55a17a641aa28af57"} Oct 13 08:57:10 crc kubenswrapper[4685]: I1013 08:57:10.925134 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-ddb2s" event={"ID":"55ccd123-f671-4230-b2d5-e6ffb265429a","Type":"ContainerStarted","Data":"081977c13ff36add4bda96a4e338d591665cbd563a9a8e467043b1d4da62eef9"} Oct 13 08:57:11 crc kubenswrapper[4685]: I1013 08:57:11.938367 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-ddb2s" event={"ID":"55ccd123-f671-4230-b2d5-e6ffb265429a","Type":"ContainerStarted","Data":"27917202f402c44553e5de85f27dce58dd1c3c8f1c5d44d991f0f79816f0dc31"} Oct 13 08:57:11 crc kubenswrapper[4685]: I1013 08:57:11.977142 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-ddb2s" podStartSLOduration=3.9771230429999997 podStartE2EDuration="3.977123043s" podCreationTimestamp="2025-10-13 08:57:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:57:11.975455807 +0000 UTC m=+757.123331578" watchObservedRunningTime="2025-10-13 08:57:11.977123043 +0000 UTC m=+757.124998804" Oct 13 08:57:12 crc kubenswrapper[4685]: I1013 08:57:12.944125 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-ddb2s" Oct 13 08:57:14 crc kubenswrapper[4685]: I1013 08:57:14.398991 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vlqs8"] Oct 13 08:57:14 crc kubenswrapper[4685]: I1013 08:57:14.400154 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vlqs8" Oct 13 08:57:14 crc kubenswrapper[4685]: I1013 08:57:14.410164 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vlqs8"] Oct 13 08:57:14 crc kubenswrapper[4685]: I1013 08:57:14.500795 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86zkc\" (UniqueName: \"kubernetes.io/projected/53a1934f-04b2-4aa8-81a7-ab0f0916371a-kube-api-access-86zkc\") pod \"community-operators-vlqs8\" (UID: \"53a1934f-04b2-4aa8-81a7-ab0f0916371a\") " pod="openshift-marketplace/community-operators-vlqs8" Oct 13 08:57:14 crc kubenswrapper[4685]: I1013 08:57:14.500878 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53a1934f-04b2-4aa8-81a7-ab0f0916371a-catalog-content\") pod \"community-operators-vlqs8\" (UID: \"53a1934f-04b2-4aa8-81a7-ab0f0916371a\") " pod="openshift-marketplace/community-operators-vlqs8" Oct 13 08:57:14 crc kubenswrapper[4685]: I1013 08:57:14.500907 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53a1934f-04b2-4aa8-81a7-ab0f0916371a-utilities\") pod \"community-operators-vlqs8\" (UID: \"53a1934f-04b2-4aa8-81a7-ab0f0916371a\") " pod="openshift-marketplace/community-operators-vlqs8" Oct 13 08:57:14 crc kubenswrapper[4685]: I1013 08:57:14.601713 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86zkc\" (UniqueName: \"kubernetes.io/projected/53a1934f-04b2-4aa8-81a7-ab0f0916371a-kube-api-access-86zkc\") pod \"community-operators-vlqs8\" (UID: \"53a1934f-04b2-4aa8-81a7-ab0f0916371a\") " pod="openshift-marketplace/community-operators-vlqs8" Oct 13 08:57:14 crc kubenswrapper[4685]: I1013 08:57:14.601765 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53a1934f-04b2-4aa8-81a7-ab0f0916371a-catalog-content\") pod \"community-operators-vlqs8\" (UID: \"53a1934f-04b2-4aa8-81a7-ab0f0916371a\") " pod="openshift-marketplace/community-operators-vlqs8" Oct 13 08:57:14 crc kubenswrapper[4685]: I1013 08:57:14.601786 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53a1934f-04b2-4aa8-81a7-ab0f0916371a-utilities\") pod \"community-operators-vlqs8\" (UID: \"53a1934f-04b2-4aa8-81a7-ab0f0916371a\") " pod="openshift-marketplace/community-operators-vlqs8" Oct 13 08:57:14 crc kubenswrapper[4685]: I1013 08:57:14.602374 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53a1934f-04b2-4aa8-81a7-ab0f0916371a-catalog-content\") pod \"community-operators-vlqs8\" (UID: \"53a1934f-04b2-4aa8-81a7-ab0f0916371a\") " pod="openshift-marketplace/community-operators-vlqs8" Oct 13 08:57:14 crc kubenswrapper[4685]: I1013 08:57:14.613041 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53a1934f-04b2-4aa8-81a7-ab0f0916371a-utilities\") pod \"community-operators-vlqs8\" (UID: \"53a1934f-04b2-4aa8-81a7-ab0f0916371a\") " pod="openshift-marketplace/community-operators-vlqs8" Oct 13 08:57:14 crc kubenswrapper[4685]: I1013 08:57:14.626221 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86zkc\" (UniqueName: \"kubernetes.io/projected/53a1934f-04b2-4aa8-81a7-ab0f0916371a-kube-api-access-86zkc\") pod \"community-operators-vlqs8\" (UID: \"53a1934f-04b2-4aa8-81a7-ab0f0916371a\") " pod="openshift-marketplace/community-operators-vlqs8" Oct 13 08:57:14 crc kubenswrapper[4685]: I1013 08:57:14.723141 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vlqs8" Oct 13 08:57:16 crc kubenswrapper[4685]: I1013 08:57:16.797995 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-q2zxc"] Oct 13 08:57:16 crc kubenswrapper[4685]: I1013 08:57:16.802530 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q2zxc" Oct 13 08:57:16 crc kubenswrapper[4685]: I1013 08:57:16.812463 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q2zxc"] Oct 13 08:57:16 crc kubenswrapper[4685]: I1013 08:57:16.830118 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e918bbfe-4b27-4013-85fb-27c1c4ea3610-utilities\") pod \"certified-operators-q2zxc\" (UID: \"e918bbfe-4b27-4013-85fb-27c1c4ea3610\") " pod="openshift-marketplace/certified-operators-q2zxc" Oct 13 08:57:16 crc kubenswrapper[4685]: I1013 08:57:16.830174 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p6lq\" (UniqueName: \"kubernetes.io/projected/e918bbfe-4b27-4013-85fb-27c1c4ea3610-kube-api-access-6p6lq\") pod \"certified-operators-q2zxc\" (UID: \"e918bbfe-4b27-4013-85fb-27c1c4ea3610\") " pod="openshift-marketplace/certified-operators-q2zxc" Oct 13 08:57:16 crc kubenswrapper[4685]: I1013 08:57:16.830234 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e918bbfe-4b27-4013-85fb-27c1c4ea3610-catalog-content\") pod \"certified-operators-q2zxc\" (UID: \"e918bbfe-4b27-4013-85fb-27c1c4ea3610\") " pod="openshift-marketplace/certified-operators-q2zxc" Oct 13 08:57:16 crc kubenswrapper[4685]: I1013 08:57:16.931773 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e918bbfe-4b27-4013-85fb-27c1c4ea3610-catalog-content\") pod \"certified-operators-q2zxc\" (UID: \"e918bbfe-4b27-4013-85fb-27c1c4ea3610\") " pod="openshift-marketplace/certified-operators-q2zxc" Oct 13 08:57:16 crc kubenswrapper[4685]: I1013 08:57:16.931830 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e918bbfe-4b27-4013-85fb-27c1c4ea3610-utilities\") pod \"certified-operators-q2zxc\" (UID: \"e918bbfe-4b27-4013-85fb-27c1c4ea3610\") " pod="openshift-marketplace/certified-operators-q2zxc" Oct 13 08:57:16 crc kubenswrapper[4685]: I1013 08:57:16.931864 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p6lq\" (UniqueName: \"kubernetes.io/projected/e918bbfe-4b27-4013-85fb-27c1c4ea3610-kube-api-access-6p6lq\") pod \"certified-operators-q2zxc\" (UID: \"e918bbfe-4b27-4013-85fb-27c1c4ea3610\") " pod="openshift-marketplace/certified-operators-q2zxc" Oct 13 08:57:16 crc kubenswrapper[4685]: I1013 08:57:16.932257 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e918bbfe-4b27-4013-85fb-27c1c4ea3610-catalog-content\") pod \"certified-operators-q2zxc\" (UID: \"e918bbfe-4b27-4013-85fb-27c1c4ea3610\") " pod="openshift-marketplace/certified-operators-q2zxc" Oct 13 08:57:16 crc kubenswrapper[4685]: I1013 08:57:16.932303 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e918bbfe-4b27-4013-85fb-27c1c4ea3610-utilities\") pod \"certified-operators-q2zxc\" (UID: \"e918bbfe-4b27-4013-85fb-27c1c4ea3610\") " pod="openshift-marketplace/certified-operators-q2zxc" Oct 13 08:57:16 crc kubenswrapper[4685]: I1013 08:57:16.948898 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p6lq\" (UniqueName: \"kubernetes.io/projected/e918bbfe-4b27-4013-85fb-27c1c4ea3610-kube-api-access-6p6lq\") pod \"certified-operators-q2zxc\" (UID: \"e918bbfe-4b27-4013-85fb-27c1c4ea3610\") " pod="openshift-marketplace/certified-operators-q2zxc" Oct 13 08:57:17 crc kubenswrapper[4685]: I1013 08:57:17.121082 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q2zxc" Oct 13 08:57:18 crc kubenswrapper[4685]: I1013 08:57:18.426011 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vlqs8"] Oct 13 08:57:18 crc kubenswrapper[4685]: W1013 08:57:18.433809 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53a1934f_04b2_4aa8_81a7_ab0f0916371a.slice/crio-51da86f391014d237cdb2e779e7cabb22419cd49466f938b1773663710ded057 WatchSource:0}: Error finding container 51da86f391014d237cdb2e779e7cabb22419cd49466f938b1773663710ded057: Status 404 returned error can't find the container with id 51da86f391014d237cdb2e779e7cabb22419cd49466f938b1773663710ded057 Oct 13 08:57:18 crc kubenswrapper[4685]: I1013 08:57:18.493597 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q2zxc"] Oct 13 08:57:18 crc kubenswrapper[4685]: I1013 08:57:18.986220 4685 generic.go:334] "Generic (PLEG): container finished" podID="e918bbfe-4b27-4013-85fb-27c1c4ea3610" containerID="970ec1c56289fe8a7c5433419155c9ad64ca78b5ed6d1b18e2d3b9b7ebf58352" exitCode=0 Oct 13 08:57:18 crc kubenswrapper[4685]: I1013 08:57:18.986283 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2zxc" event={"ID":"e918bbfe-4b27-4013-85fb-27c1c4ea3610","Type":"ContainerDied","Data":"970ec1c56289fe8a7c5433419155c9ad64ca78b5ed6d1b18e2d3b9b7ebf58352"} Oct 13 08:57:18 crc kubenswrapper[4685]: I1013 08:57:18.986364 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2zxc" event={"ID":"e918bbfe-4b27-4013-85fb-27c1c4ea3610","Type":"ContainerStarted","Data":"1c8d3fb0178e7b52b15cb670ac31ad9f937d0c945255a9f4c2690438c6d69520"} Oct 13 08:57:18 crc kubenswrapper[4685]: I1013 08:57:18.988874 4685 generic.go:334] "Generic (PLEG): container finished" podID="33a86974-4a79-42b8-8bf9-cf3992b9b54d" containerID="c8303d80d1f9364d5d01a83483d72486cb0effd753c4251a57aa125978bf0da0" exitCode=0 Oct 13 08:57:18 crc kubenswrapper[4685]: I1013 08:57:18.988990 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ggf8f" event={"ID":"33a86974-4a79-42b8-8bf9-cf3992b9b54d","Type":"ContainerDied","Data":"c8303d80d1f9364d5d01a83483d72486cb0effd753c4251a57aa125978bf0da0"} Oct 13 08:57:18 crc kubenswrapper[4685]: I1013 08:57:18.991430 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-fkd9z" event={"ID":"9b097f6c-ec81-4342-b4c1-63520267ba08","Type":"ContainerStarted","Data":"9a1523782d9630d0afbaefdabb2e80ad1a053deb034cf7db4f4271d3495b9e75"} Oct 13 08:57:18 crc kubenswrapper[4685]: I1013 08:57:18.991595 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-fkd9z" Oct 13 08:57:18 crc kubenswrapper[4685]: I1013 08:57:18.993296 4685 generic.go:334] "Generic (PLEG): container finished" podID="53a1934f-04b2-4aa8-81a7-ab0f0916371a" containerID="8955b2c0d20b908eb1f7f3436747ebaa4031255c80e971eecf73046caca21b0f" exitCode=0 Oct 13 08:57:18 crc kubenswrapper[4685]: I1013 08:57:18.993354 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vlqs8" event={"ID":"53a1934f-04b2-4aa8-81a7-ab0f0916371a","Type":"ContainerDied","Data":"8955b2c0d20b908eb1f7f3436747ebaa4031255c80e971eecf73046caca21b0f"} Oct 13 08:57:18 crc kubenswrapper[4685]: I1013 08:57:18.993430 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vlqs8" event={"ID":"53a1934f-04b2-4aa8-81a7-ab0f0916371a","Type":"ContainerStarted","Data":"51da86f391014d237cdb2e779e7cabb22419cd49466f938b1773663710ded057"} Oct 13 08:57:19 crc kubenswrapper[4685]: I1013 08:57:19.051675 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-fkd9z" podStartSLOduration=2.703499041 podStartE2EDuration="11.051658303s" podCreationTimestamp="2025-10-13 08:57:08 +0000 UTC" firstStartedPulling="2025-10-13 08:57:09.699112141 +0000 UTC m=+754.846987902" lastFinishedPulling="2025-10-13 08:57:18.047271403 +0000 UTC m=+763.195147164" observedRunningTime="2025-10-13 08:57:19.048811495 +0000 UTC m=+764.196687266" watchObservedRunningTime="2025-10-13 08:57:19.051658303 +0000 UTC m=+764.199534074" Oct 13 08:57:20 crc kubenswrapper[4685]: I1013 08:57:20.003857 4685 generic.go:334] "Generic (PLEG): container finished" podID="33a86974-4a79-42b8-8bf9-cf3992b9b54d" containerID="920970ed490770b0a6d77a509e801b3685ccc7a51c65bacd722e437bb5496118" exitCode=0 Oct 13 08:57:20 crc kubenswrapper[4685]: I1013 08:57:20.003947 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ggf8f" event={"ID":"33a86974-4a79-42b8-8bf9-cf3992b9b54d","Type":"ContainerDied","Data":"920970ed490770b0a6d77a509e801b3685ccc7a51c65bacd722e437bb5496118"} Oct 13 08:57:20 crc kubenswrapper[4685]: I1013 08:57:20.010818 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2zxc" event={"ID":"e918bbfe-4b27-4013-85fb-27c1c4ea3610","Type":"ContainerStarted","Data":"4d13e0190455f17ad7cd8296b8d07a1af23bac0821ce05de3d0631f0110ffef1"} Oct 13 08:57:20 crc kubenswrapper[4685]: I1013 08:57:20.343740 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-ddb2s" Oct 13 08:57:21 crc kubenswrapper[4685]: I1013 08:57:21.023312 4685 generic.go:334] "Generic (PLEG): container finished" podID="e918bbfe-4b27-4013-85fb-27c1c4ea3610" containerID="4d13e0190455f17ad7cd8296b8d07a1af23bac0821ce05de3d0631f0110ffef1" exitCode=0 Oct 13 08:57:21 crc kubenswrapper[4685]: I1013 08:57:21.023361 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2zxc" event={"ID":"e918bbfe-4b27-4013-85fb-27c1c4ea3610","Type":"ContainerDied","Data":"4d13e0190455f17ad7cd8296b8d07a1af23bac0821ce05de3d0631f0110ffef1"} Oct 13 08:57:21 crc kubenswrapper[4685]: I1013 08:57:21.028314 4685 generic.go:334] "Generic (PLEG): container finished" podID="33a86974-4a79-42b8-8bf9-cf3992b9b54d" containerID="e80a525c8d2d84438d6b6e72239afbbc7e1d93605f40020479e180d6b13df935" exitCode=0 Oct 13 08:57:21 crc kubenswrapper[4685]: I1013 08:57:21.028653 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ggf8f" event={"ID":"33a86974-4a79-42b8-8bf9-cf3992b9b54d","Type":"ContainerDied","Data":"e80a525c8d2d84438d6b6e72239afbbc7e1d93605f40020479e180d6b13df935"} Oct 13 08:57:21 crc kubenswrapper[4685]: I1013 08:57:21.034797 4685 generic.go:334] "Generic (PLEG): container finished" podID="53a1934f-04b2-4aa8-81a7-ab0f0916371a" containerID="a8dd4dec6a4cde97ae74dc410331d9839f55d6d91ecf122c89d8345f5ad4d5db" exitCode=0 Oct 13 08:57:21 crc kubenswrapper[4685]: I1013 08:57:21.034882 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vlqs8" event={"ID":"53a1934f-04b2-4aa8-81a7-ab0f0916371a","Type":"ContainerDied","Data":"a8dd4dec6a4cde97ae74dc410331d9839f55d6d91ecf122c89d8345f5ad4d5db"} Oct 13 08:57:22 crc kubenswrapper[4685]: I1013 08:57:22.055006 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ggf8f" event={"ID":"33a86974-4a79-42b8-8bf9-cf3992b9b54d","Type":"ContainerStarted","Data":"8f18a19bd93e082f446254f377b338116f98092e3392696d4d105c023a1ecdfa"} Oct 13 08:57:22 crc kubenswrapper[4685]: I1013 08:57:22.055251 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ggf8f" event={"ID":"33a86974-4a79-42b8-8bf9-cf3992b9b54d","Type":"ContainerStarted","Data":"49f5f54d42279121f3abc53de21a12970ddb6a2c856d5e9b1746337e113bef7c"} Oct 13 08:57:22 crc kubenswrapper[4685]: I1013 08:57:22.055262 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ggf8f" event={"ID":"33a86974-4a79-42b8-8bf9-cf3992b9b54d","Type":"ContainerStarted","Data":"32b47df2c31ceeca3fae82e31288a786a5277fd3fbc929960ce9485ab20f9863"} Oct 13 08:57:22 crc kubenswrapper[4685]: I1013 08:57:22.055271 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ggf8f" event={"ID":"33a86974-4a79-42b8-8bf9-cf3992b9b54d","Type":"ContainerStarted","Data":"0f29e2ed2fe35e78bf441050371ca8ebc07a464fa8f96582bdc57d4565aa1a43"} Oct 13 08:57:22 crc kubenswrapper[4685]: I1013 08:57:22.055279 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ggf8f" event={"ID":"33a86974-4a79-42b8-8bf9-cf3992b9b54d","Type":"ContainerStarted","Data":"a0123ccd058a3e82d46b9ef1e664da7ad2317cc493cc63ba1fc84df605cc5074"} Oct 13 08:57:22 crc kubenswrapper[4685]: I1013 08:57:22.062182 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2zxc" event={"ID":"e918bbfe-4b27-4013-85fb-27c1c4ea3610","Type":"ContainerStarted","Data":"18e80c4e38c4755f704712979400226515bbb36cdac1a7a15c3998a40615c0e9"} Oct 13 08:57:22 crc kubenswrapper[4685]: I1013 08:57:22.067008 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vlqs8" event={"ID":"53a1934f-04b2-4aa8-81a7-ab0f0916371a","Type":"ContainerStarted","Data":"78982279b3ca69273407446528e57e143c445bb076044028852c1293bc818491"} Oct 13 08:57:22 crc kubenswrapper[4685]: I1013 08:57:22.118423 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vlqs8" podStartSLOduration=5.647955193 podStartE2EDuration="8.118409802s" podCreationTimestamp="2025-10-13 08:57:14 +0000 UTC" firstStartedPulling="2025-10-13 08:57:18.994744952 +0000 UTC m=+764.142620723" lastFinishedPulling="2025-10-13 08:57:21.465199571 +0000 UTC m=+766.613075332" observedRunningTime="2025-10-13 08:57:22.115352338 +0000 UTC m=+767.263228089" watchObservedRunningTime="2025-10-13 08:57:22.118409802 +0000 UTC m=+767.266285553" Oct 13 08:57:22 crc kubenswrapper[4685]: I1013 08:57:22.119724 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-q2zxc" podStartSLOduration=3.625046704 podStartE2EDuration="6.119715207s" podCreationTimestamp="2025-10-13 08:57:16 +0000 UTC" firstStartedPulling="2025-10-13 08:57:18.98920568 +0000 UTC m=+764.137081481" lastFinishedPulling="2025-10-13 08:57:21.483874233 +0000 UTC m=+766.631749984" observedRunningTime="2025-10-13 08:57:22.09099732 +0000 UTC m=+767.238873081" watchObservedRunningTime="2025-10-13 08:57:22.119715207 +0000 UTC m=+767.267590968" Oct 13 08:57:23 crc kubenswrapper[4685]: I1013 08:57:23.078357 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ggf8f" event={"ID":"33a86974-4a79-42b8-8bf9-cf3992b9b54d","Type":"ContainerStarted","Data":"d12bb0064394a329463ed15bb53f3684013017724ac6f6c878b8919825b7fb18"} Oct 13 08:57:23 crc kubenswrapper[4685]: I1013 08:57:23.118525 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-ggf8f" podStartSLOduration=5.932999162 podStartE2EDuration="15.118499014s" podCreationTimestamp="2025-10-13 08:57:08 +0000 UTC" firstStartedPulling="2025-10-13 08:57:08.848089436 +0000 UTC m=+753.995965197" lastFinishedPulling="2025-10-13 08:57:18.033589298 +0000 UTC m=+763.181465049" observedRunningTime="2025-10-13 08:57:23.114397021 +0000 UTC m=+768.262272822" watchObservedRunningTime="2025-10-13 08:57:23.118499014 +0000 UTC m=+768.266374785" Oct 13 08:57:23 crc kubenswrapper[4685]: I1013 08:57:23.678306 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:23 crc kubenswrapper[4685]: I1013 08:57:23.721639 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:24 crc kubenswrapper[4685]: I1013 08:57:24.083948 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:24 crc kubenswrapper[4685]: I1013 08:57:24.725060 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vlqs8" Oct 13 08:57:24 crc kubenswrapper[4685]: I1013 08:57:24.725587 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vlqs8" Oct 13 08:57:24 crc kubenswrapper[4685]: I1013 08:57:24.774256 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vlqs8" Oct 13 08:57:27 crc kubenswrapper[4685]: I1013 08:57:27.121742 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-q2zxc" Oct 13 08:57:27 crc kubenswrapper[4685]: I1013 08:57:27.122225 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-q2zxc" Oct 13 08:57:27 crc kubenswrapper[4685]: I1013 08:57:27.194084 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-q2zxc" Oct 13 08:57:28 crc kubenswrapper[4685]: I1013 08:57:28.158277 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-q2zxc" Oct 13 08:57:28 crc kubenswrapper[4685]: I1013 08:57:28.864664 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-fmsz9" Oct 13 08:57:29 crc kubenswrapper[4685]: I1013 08:57:29.176970 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-t7wsk"] Oct 13 08:57:29 crc kubenswrapper[4685]: I1013 08:57:29.177847 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-t7wsk" Oct 13 08:57:29 crc kubenswrapper[4685]: I1013 08:57:29.180589 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 13 08:57:29 crc kubenswrapper[4685]: I1013 08:57:29.182040 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 13 08:57:29 crc kubenswrapper[4685]: I1013 08:57:29.186481 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-h2xbz" Oct 13 08:57:29 crc kubenswrapper[4685]: I1013 08:57:29.191391 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-t7wsk"] Oct 13 08:57:29 crc kubenswrapper[4685]: I1013 08:57:29.303739 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsjzx\" (UniqueName: \"kubernetes.io/projected/3d6866fd-e05d-4533-9134-47a83469940b-kube-api-access-jsjzx\") pod \"openstack-operator-index-t7wsk\" (UID: \"3d6866fd-e05d-4533-9134-47a83469940b\") " pod="openstack-operators/openstack-operator-index-t7wsk" Oct 13 08:57:29 crc kubenswrapper[4685]: I1013 08:57:29.307885 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-fkd9z" Oct 13 08:57:29 crc kubenswrapper[4685]: I1013 08:57:29.405610 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsjzx\" (UniqueName: \"kubernetes.io/projected/3d6866fd-e05d-4533-9134-47a83469940b-kube-api-access-jsjzx\") pod \"openstack-operator-index-t7wsk\" (UID: \"3d6866fd-e05d-4533-9134-47a83469940b\") " pod="openstack-operators/openstack-operator-index-t7wsk" Oct 13 08:57:29 crc kubenswrapper[4685]: I1013 08:57:29.424302 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsjzx\" (UniqueName: \"kubernetes.io/projected/3d6866fd-e05d-4533-9134-47a83469940b-kube-api-access-jsjzx\") pod \"openstack-operator-index-t7wsk\" (UID: \"3d6866fd-e05d-4533-9134-47a83469940b\") " pod="openstack-operators/openstack-operator-index-t7wsk" Oct 13 08:57:29 crc kubenswrapper[4685]: I1013 08:57:29.505553 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-t7wsk" Oct 13 08:57:30 crc kubenswrapper[4685]: I1013 08:57:30.510640 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-t7wsk"] Oct 13 08:57:31 crc kubenswrapper[4685]: I1013 08:57:31.135895 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-t7wsk" event={"ID":"3d6866fd-e05d-4533-9134-47a83469940b","Type":"ContainerStarted","Data":"9d393362e0f45ebae783f6f8458dcc86a382d60e5b3f6ca2d3e07be485aae6e7"} Oct 13 08:57:33 crc kubenswrapper[4685]: I1013 08:57:33.149668 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-t7wsk" event={"ID":"3d6866fd-e05d-4533-9134-47a83469940b","Type":"ContainerStarted","Data":"68509516fdb2129c28253367bd3d1f9fc766bd521d26f3f3f96f6d11eedefaf3"} Oct 13 08:57:33 crc kubenswrapper[4685]: I1013 08:57:33.770215 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-t7wsk" podStartSLOduration=2.842403388 podStartE2EDuration="4.770189977s" podCreationTimestamp="2025-10-13 08:57:29 +0000 UTC" firstStartedPulling="2025-10-13 08:57:30.528016849 +0000 UTC m=+775.675892610" lastFinishedPulling="2025-10-13 08:57:32.455803428 +0000 UTC m=+777.603679199" observedRunningTime="2025-10-13 08:57:33.172267113 +0000 UTC m=+778.320142914" watchObservedRunningTime="2025-10-13 08:57:33.770189977 +0000 UTC m=+778.918065778" Oct 13 08:57:33 crc kubenswrapper[4685]: I1013 08:57:33.772254 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q2zxc"] Oct 13 08:57:33 crc kubenswrapper[4685]: I1013 08:57:33.772588 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-q2zxc" podUID="e918bbfe-4b27-4013-85fb-27c1c4ea3610" containerName="registry-server" containerID="cri-o://18e80c4e38c4755f704712979400226515bbb36cdac1a7a15c3998a40615c0e9" gracePeriod=2 Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.160622 4685 generic.go:334] "Generic (PLEG): container finished" podID="e918bbfe-4b27-4013-85fb-27c1c4ea3610" containerID="18e80c4e38c4755f704712979400226515bbb36cdac1a7a15c3998a40615c0e9" exitCode=0 Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.160752 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2zxc" event={"ID":"e918bbfe-4b27-4013-85fb-27c1c4ea3610","Type":"ContainerDied","Data":"18e80c4e38c4755f704712979400226515bbb36cdac1a7a15c3998a40615c0e9"} Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.230443 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-z9vp4"] Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.231820 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z9vp4" Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.246531 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z9vp4"] Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.270837 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngkfz\" (UniqueName: \"kubernetes.io/projected/0141650e-9da4-48d3-a6d8-d0d238397f97-kube-api-access-ngkfz\") pod \"redhat-marketplace-z9vp4\" (UID: \"0141650e-9da4-48d3-a6d8-d0d238397f97\") " pod="openshift-marketplace/redhat-marketplace-z9vp4" Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.270897 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0141650e-9da4-48d3-a6d8-d0d238397f97-catalog-content\") pod \"redhat-marketplace-z9vp4\" (UID: \"0141650e-9da4-48d3-a6d8-d0d238397f97\") " pod="openshift-marketplace/redhat-marketplace-z9vp4" Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.270952 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0141650e-9da4-48d3-a6d8-d0d238397f97-utilities\") pod \"redhat-marketplace-z9vp4\" (UID: \"0141650e-9da4-48d3-a6d8-d0d238397f97\") " pod="openshift-marketplace/redhat-marketplace-z9vp4" Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.371737 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngkfz\" (UniqueName: \"kubernetes.io/projected/0141650e-9da4-48d3-a6d8-d0d238397f97-kube-api-access-ngkfz\") pod \"redhat-marketplace-z9vp4\" (UID: \"0141650e-9da4-48d3-a6d8-d0d238397f97\") " pod="openshift-marketplace/redhat-marketplace-z9vp4" Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.371785 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0141650e-9da4-48d3-a6d8-d0d238397f97-catalog-content\") pod \"redhat-marketplace-z9vp4\" (UID: \"0141650e-9da4-48d3-a6d8-d0d238397f97\") " pod="openshift-marketplace/redhat-marketplace-z9vp4" Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.371820 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0141650e-9da4-48d3-a6d8-d0d238397f97-utilities\") pod \"redhat-marketplace-z9vp4\" (UID: \"0141650e-9da4-48d3-a6d8-d0d238397f97\") " pod="openshift-marketplace/redhat-marketplace-z9vp4" Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.372528 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0141650e-9da4-48d3-a6d8-d0d238397f97-catalog-content\") pod \"redhat-marketplace-z9vp4\" (UID: \"0141650e-9da4-48d3-a6d8-d0d238397f97\") " pod="openshift-marketplace/redhat-marketplace-z9vp4" Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.372851 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0141650e-9da4-48d3-a6d8-d0d238397f97-utilities\") pod \"redhat-marketplace-z9vp4\" (UID: \"0141650e-9da4-48d3-a6d8-d0d238397f97\") " pod="openshift-marketplace/redhat-marketplace-z9vp4" Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.374467 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q2zxc" Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.406260 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngkfz\" (UniqueName: \"kubernetes.io/projected/0141650e-9da4-48d3-a6d8-d0d238397f97-kube-api-access-ngkfz\") pod \"redhat-marketplace-z9vp4\" (UID: \"0141650e-9da4-48d3-a6d8-d0d238397f97\") " pod="openshift-marketplace/redhat-marketplace-z9vp4" Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.473090 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e918bbfe-4b27-4013-85fb-27c1c4ea3610-catalog-content\") pod \"e918bbfe-4b27-4013-85fb-27c1c4ea3610\" (UID: \"e918bbfe-4b27-4013-85fb-27c1c4ea3610\") " Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.473177 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e918bbfe-4b27-4013-85fb-27c1c4ea3610-utilities\") pod \"e918bbfe-4b27-4013-85fb-27c1c4ea3610\" (UID: \"e918bbfe-4b27-4013-85fb-27c1c4ea3610\") " Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.473275 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6p6lq\" (UniqueName: \"kubernetes.io/projected/e918bbfe-4b27-4013-85fb-27c1c4ea3610-kube-api-access-6p6lq\") pod \"e918bbfe-4b27-4013-85fb-27c1c4ea3610\" (UID: \"e918bbfe-4b27-4013-85fb-27c1c4ea3610\") " Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.474085 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e918bbfe-4b27-4013-85fb-27c1c4ea3610-utilities" (OuterVolumeSpecName: "utilities") pod "e918bbfe-4b27-4013-85fb-27c1c4ea3610" (UID: "e918bbfe-4b27-4013-85fb-27c1c4ea3610"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.487029 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e918bbfe-4b27-4013-85fb-27c1c4ea3610-kube-api-access-6p6lq" (OuterVolumeSpecName: "kube-api-access-6p6lq") pod "e918bbfe-4b27-4013-85fb-27c1c4ea3610" (UID: "e918bbfe-4b27-4013-85fb-27c1c4ea3610"). InnerVolumeSpecName "kube-api-access-6p6lq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.521271 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e918bbfe-4b27-4013-85fb-27c1c4ea3610-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e918bbfe-4b27-4013-85fb-27c1c4ea3610" (UID: "e918bbfe-4b27-4013-85fb-27c1c4ea3610"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.552185 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z9vp4" Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.574495 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6p6lq\" (UniqueName: \"kubernetes.io/projected/e918bbfe-4b27-4013-85fb-27c1c4ea3610-kube-api-access-6p6lq\") on node \"crc\" DevicePath \"\"" Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.574801 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e918bbfe-4b27-4013-85fb-27c1c4ea3610-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.574811 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e918bbfe-4b27-4013-85fb-27c1c4ea3610-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 08:57:34 crc kubenswrapper[4685]: I1013 08:57:34.771475 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vlqs8" Oct 13 08:57:35 crc kubenswrapper[4685]: I1013 08:57:35.001399 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z9vp4"] Oct 13 08:57:35 crc kubenswrapper[4685]: W1013 08:57:35.010199 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0141650e_9da4_48d3_a6d8_d0d238397f97.slice/crio-f86165a416e35b0c41ff049d9f631942c0c6ba02550f040ef4e2035122fa4ab4 WatchSource:0}: Error finding container f86165a416e35b0c41ff049d9f631942c0c6ba02550f040ef4e2035122fa4ab4: Status 404 returned error can't find the container with id f86165a416e35b0c41ff049d9f631942c0c6ba02550f040ef4e2035122fa4ab4 Oct 13 08:57:35 crc kubenswrapper[4685]: I1013 08:57:35.185501 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2zxc" event={"ID":"e918bbfe-4b27-4013-85fb-27c1c4ea3610","Type":"ContainerDied","Data":"1c8d3fb0178e7b52b15cb670ac31ad9f937d0c945255a9f4c2690438c6d69520"} Oct 13 08:57:35 crc kubenswrapper[4685]: I1013 08:57:35.185878 4685 scope.go:117] "RemoveContainer" containerID="18e80c4e38c4755f704712979400226515bbb36cdac1a7a15c3998a40615c0e9" Oct 13 08:57:35 crc kubenswrapper[4685]: I1013 08:57:35.185680 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q2zxc" Oct 13 08:57:35 crc kubenswrapper[4685]: I1013 08:57:35.188669 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z9vp4" event={"ID":"0141650e-9da4-48d3-a6d8-d0d238397f97","Type":"ContainerStarted","Data":"f86165a416e35b0c41ff049d9f631942c0c6ba02550f040ef4e2035122fa4ab4"} Oct 13 08:57:35 crc kubenswrapper[4685]: I1013 08:57:35.207685 4685 scope.go:117] "RemoveContainer" containerID="4d13e0190455f17ad7cd8296b8d07a1af23bac0821ce05de3d0631f0110ffef1" Oct 13 08:57:35 crc kubenswrapper[4685]: I1013 08:57:35.224666 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q2zxc"] Oct 13 08:57:35 crc kubenswrapper[4685]: I1013 08:57:35.228695 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-q2zxc"] Oct 13 08:57:35 crc kubenswrapper[4685]: I1013 08:57:35.233067 4685 scope.go:117] "RemoveContainer" containerID="970ec1c56289fe8a7c5433419155c9ad64ca78b5ed6d1b18e2d3b9b7ebf58352" Oct 13 08:57:35 crc kubenswrapper[4685]: I1013 08:57:35.537480 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e918bbfe-4b27-4013-85fb-27c1c4ea3610" path="/var/lib/kubelet/pods/e918bbfe-4b27-4013-85fb-27c1c4ea3610/volumes" Oct 13 08:57:36 crc kubenswrapper[4685]: I1013 08:57:36.195848 4685 generic.go:334] "Generic (PLEG): container finished" podID="0141650e-9da4-48d3-a6d8-d0d238397f97" containerID="ba32490fa78c0658c7c8915bd33348dd97becd0d2a423a86742e7e8ee2431145" exitCode=0 Oct 13 08:57:36 crc kubenswrapper[4685]: I1013 08:57:36.195896 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z9vp4" event={"ID":"0141650e-9da4-48d3-a6d8-d0d238397f97","Type":"ContainerDied","Data":"ba32490fa78c0658c7c8915bd33348dd97becd0d2a423a86742e7e8ee2431145"} Oct 13 08:57:38 crc kubenswrapper[4685]: I1013 08:57:38.209111 4685 generic.go:334] "Generic (PLEG): container finished" podID="0141650e-9da4-48d3-a6d8-d0d238397f97" containerID="3e13a94a7f5d614150cc45cd511c77e8f4f2c6f1fd4003bed7423804f4f73b02" exitCode=0 Oct 13 08:57:38 crc kubenswrapper[4685]: I1013 08:57:38.209165 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z9vp4" event={"ID":"0141650e-9da4-48d3-a6d8-d0d238397f97","Type":"ContainerDied","Data":"3e13a94a7f5d614150cc45cd511c77e8f4f2c6f1fd4003bed7423804f4f73b02"} Oct 13 08:57:38 crc kubenswrapper[4685]: I1013 08:57:38.680075 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-ggf8f" Oct 13 08:57:39 crc kubenswrapper[4685]: I1013 08:57:39.216585 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z9vp4" event={"ID":"0141650e-9da4-48d3-a6d8-d0d238397f97","Type":"ContainerStarted","Data":"47d5d87689bea2586fa6f54967b4a9e7115593808d1ee38bad3dab66934544f3"} Oct 13 08:57:39 crc kubenswrapper[4685]: I1013 08:57:39.237350 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-z9vp4" podStartSLOduration=2.6599952350000002 podStartE2EDuration="5.237325574s" podCreationTimestamp="2025-10-13 08:57:34 +0000 UTC" firstStartedPulling="2025-10-13 08:57:36.197498794 +0000 UTC m=+781.345374555" lastFinishedPulling="2025-10-13 08:57:38.774829123 +0000 UTC m=+783.922704894" observedRunningTime="2025-10-13 08:57:39.236514561 +0000 UTC m=+784.384390342" watchObservedRunningTime="2025-10-13 08:57:39.237325574 +0000 UTC m=+784.385201345" Oct 13 08:57:39 crc kubenswrapper[4685]: I1013 08:57:39.515546 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-t7wsk" Oct 13 08:57:39 crc kubenswrapper[4685]: I1013 08:57:39.515907 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-t7wsk" Oct 13 08:57:39 crc kubenswrapper[4685]: I1013 08:57:39.540636 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-t7wsk" Oct 13 08:57:40 crc kubenswrapper[4685]: I1013 08:57:40.247304 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-t7wsk" Oct 13 08:57:40 crc kubenswrapper[4685]: I1013 08:57:40.967991 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vlqs8"] Oct 13 08:57:40 crc kubenswrapper[4685]: I1013 08:57:40.968240 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vlqs8" podUID="53a1934f-04b2-4aa8-81a7-ab0f0916371a" containerName="registry-server" containerID="cri-o://78982279b3ca69273407446528e57e143c445bb076044028852c1293bc818491" gracePeriod=2 Oct 13 08:57:41 crc kubenswrapper[4685]: I1013 08:57:41.231487 4685 generic.go:334] "Generic (PLEG): container finished" podID="53a1934f-04b2-4aa8-81a7-ab0f0916371a" containerID="78982279b3ca69273407446528e57e143c445bb076044028852c1293bc818491" exitCode=0 Oct 13 08:57:41 crc kubenswrapper[4685]: I1013 08:57:41.231770 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vlqs8" event={"ID":"53a1934f-04b2-4aa8-81a7-ab0f0916371a","Type":"ContainerDied","Data":"78982279b3ca69273407446528e57e143c445bb076044028852c1293bc818491"} Oct 13 08:57:41 crc kubenswrapper[4685]: I1013 08:57:41.460011 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vlqs8" Oct 13 08:57:41 crc kubenswrapper[4685]: I1013 08:57:41.563422 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53a1934f-04b2-4aa8-81a7-ab0f0916371a-utilities\") pod \"53a1934f-04b2-4aa8-81a7-ab0f0916371a\" (UID: \"53a1934f-04b2-4aa8-81a7-ab0f0916371a\") " Oct 13 08:57:41 crc kubenswrapper[4685]: I1013 08:57:41.563466 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53a1934f-04b2-4aa8-81a7-ab0f0916371a-catalog-content\") pod \"53a1934f-04b2-4aa8-81a7-ab0f0916371a\" (UID: \"53a1934f-04b2-4aa8-81a7-ab0f0916371a\") " Oct 13 08:57:41 crc kubenswrapper[4685]: I1013 08:57:41.563599 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86zkc\" (UniqueName: \"kubernetes.io/projected/53a1934f-04b2-4aa8-81a7-ab0f0916371a-kube-api-access-86zkc\") pod \"53a1934f-04b2-4aa8-81a7-ab0f0916371a\" (UID: \"53a1934f-04b2-4aa8-81a7-ab0f0916371a\") " Oct 13 08:57:41 crc kubenswrapper[4685]: I1013 08:57:41.565135 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53a1934f-04b2-4aa8-81a7-ab0f0916371a-utilities" (OuterVolumeSpecName: "utilities") pod "53a1934f-04b2-4aa8-81a7-ab0f0916371a" (UID: "53a1934f-04b2-4aa8-81a7-ab0f0916371a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:57:41 crc kubenswrapper[4685]: I1013 08:57:41.573101 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53a1934f-04b2-4aa8-81a7-ab0f0916371a-kube-api-access-86zkc" (OuterVolumeSpecName: "kube-api-access-86zkc") pod "53a1934f-04b2-4aa8-81a7-ab0f0916371a" (UID: "53a1934f-04b2-4aa8-81a7-ab0f0916371a"). InnerVolumeSpecName "kube-api-access-86zkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:57:41 crc kubenswrapper[4685]: I1013 08:57:41.660988 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53a1934f-04b2-4aa8-81a7-ab0f0916371a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "53a1934f-04b2-4aa8-81a7-ab0f0916371a" (UID: "53a1934f-04b2-4aa8-81a7-ab0f0916371a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:57:41 crc kubenswrapper[4685]: I1013 08:57:41.665544 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86zkc\" (UniqueName: \"kubernetes.io/projected/53a1934f-04b2-4aa8-81a7-ab0f0916371a-kube-api-access-86zkc\") on node \"crc\" DevicePath \"\"" Oct 13 08:57:41 crc kubenswrapper[4685]: I1013 08:57:41.665576 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53a1934f-04b2-4aa8-81a7-ab0f0916371a-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 08:57:41 crc kubenswrapper[4685]: I1013 08:57:41.665586 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53a1934f-04b2-4aa8-81a7-ab0f0916371a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 08:57:42 crc kubenswrapper[4685]: I1013 08:57:42.241846 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vlqs8" Oct 13 08:57:42 crc kubenswrapper[4685]: I1013 08:57:42.242411 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vlqs8" event={"ID":"53a1934f-04b2-4aa8-81a7-ab0f0916371a","Type":"ContainerDied","Data":"51da86f391014d237cdb2e779e7cabb22419cd49466f938b1773663710ded057"} Oct 13 08:57:42 crc kubenswrapper[4685]: I1013 08:57:42.242498 4685 scope.go:117] "RemoveContainer" containerID="78982279b3ca69273407446528e57e143c445bb076044028852c1293bc818491" Oct 13 08:57:42 crc kubenswrapper[4685]: I1013 08:57:42.267665 4685 scope.go:117] "RemoveContainer" containerID="a8dd4dec6a4cde97ae74dc410331d9839f55d6d91ecf122c89d8345f5ad4d5db" Oct 13 08:57:42 crc kubenswrapper[4685]: I1013 08:57:42.295666 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vlqs8"] Oct 13 08:57:42 crc kubenswrapper[4685]: I1013 08:57:42.299292 4685 scope.go:117] "RemoveContainer" containerID="8955b2c0d20b908eb1f7f3436747ebaa4031255c80e971eecf73046caca21b0f" Oct 13 08:57:42 crc kubenswrapper[4685]: I1013 08:57:42.303360 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vlqs8"] Oct 13 08:57:43 crc kubenswrapper[4685]: I1013 08:57:43.511182 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53a1934f-04b2-4aa8-81a7-ab0f0916371a" path="/var/lib/kubelet/pods/53a1934f-04b2-4aa8-81a7-ab0f0916371a/volumes" Oct 13 08:57:44 crc kubenswrapper[4685]: I1013 08:57:44.553244 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-z9vp4" Oct 13 08:57:44 crc kubenswrapper[4685]: I1013 08:57:44.553296 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-z9vp4" Oct 13 08:57:44 crc kubenswrapper[4685]: I1013 08:57:44.588871 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-z9vp4" Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.229191 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56"] Oct 13 08:57:45 crc kubenswrapper[4685]: E1013 08:57:45.229595 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e918bbfe-4b27-4013-85fb-27c1c4ea3610" containerName="extract-content" Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.229636 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e918bbfe-4b27-4013-85fb-27c1c4ea3610" containerName="extract-content" Oct 13 08:57:45 crc kubenswrapper[4685]: E1013 08:57:45.229664 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53a1934f-04b2-4aa8-81a7-ab0f0916371a" containerName="registry-server" Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.229680 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="53a1934f-04b2-4aa8-81a7-ab0f0916371a" containerName="registry-server" Oct 13 08:57:45 crc kubenswrapper[4685]: E1013 08:57:45.229711 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e918bbfe-4b27-4013-85fb-27c1c4ea3610" containerName="registry-server" Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.229730 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e918bbfe-4b27-4013-85fb-27c1c4ea3610" containerName="registry-server" Oct 13 08:57:45 crc kubenswrapper[4685]: E1013 08:57:45.229753 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53a1934f-04b2-4aa8-81a7-ab0f0916371a" containerName="extract-content" Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.229770 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="53a1934f-04b2-4aa8-81a7-ab0f0916371a" containerName="extract-content" Oct 13 08:57:45 crc kubenswrapper[4685]: E1013 08:57:45.229797 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e918bbfe-4b27-4013-85fb-27c1c4ea3610" containerName="extract-utilities" Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.229811 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e918bbfe-4b27-4013-85fb-27c1c4ea3610" containerName="extract-utilities" Oct 13 08:57:45 crc kubenswrapper[4685]: E1013 08:57:45.229844 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53a1934f-04b2-4aa8-81a7-ab0f0916371a" containerName="extract-utilities" Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.229864 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="53a1934f-04b2-4aa8-81a7-ab0f0916371a" containerName="extract-utilities" Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.230152 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="e918bbfe-4b27-4013-85fb-27c1c4ea3610" containerName="registry-server" Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.230185 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="53a1934f-04b2-4aa8-81a7-ab0f0916371a" containerName="registry-server" Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.232401 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56" Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.238402 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56"] Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.239108 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-5bspc" Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.313455 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-z9vp4" Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.315964 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c967b8a-ec86-413b-a25e-b81f3e9ac2dc-util\") pod \"4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56\" (UID: \"6c967b8a-ec86-413b-a25e-b81f3e9ac2dc\") " pod="openstack-operators/4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56" Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.316151 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2q9f\" (UniqueName: \"kubernetes.io/projected/6c967b8a-ec86-413b-a25e-b81f3e9ac2dc-kube-api-access-r2q9f\") pod \"4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56\" (UID: \"6c967b8a-ec86-413b-a25e-b81f3e9ac2dc\") " pod="openstack-operators/4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56" Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.316282 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c967b8a-ec86-413b-a25e-b81f3e9ac2dc-bundle\") pod \"4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56\" (UID: \"6c967b8a-ec86-413b-a25e-b81f3e9ac2dc\") " pod="openstack-operators/4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56" Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.417708 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c967b8a-ec86-413b-a25e-b81f3e9ac2dc-util\") pod \"4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56\" (UID: \"6c967b8a-ec86-413b-a25e-b81f3e9ac2dc\") " pod="openstack-operators/4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56" Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.418167 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2q9f\" (UniqueName: \"kubernetes.io/projected/6c967b8a-ec86-413b-a25e-b81f3e9ac2dc-kube-api-access-r2q9f\") pod \"4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56\" (UID: \"6c967b8a-ec86-413b-a25e-b81f3e9ac2dc\") " pod="openstack-operators/4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56" Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.418196 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c967b8a-ec86-413b-a25e-b81f3e9ac2dc-bundle\") pod \"4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56\" (UID: \"6c967b8a-ec86-413b-a25e-b81f3e9ac2dc\") " pod="openstack-operators/4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56" Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.418522 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c967b8a-ec86-413b-a25e-b81f3e9ac2dc-bundle\") pod \"4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56\" (UID: \"6c967b8a-ec86-413b-a25e-b81f3e9ac2dc\") " pod="openstack-operators/4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56" Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.418676 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c967b8a-ec86-413b-a25e-b81f3e9ac2dc-util\") pod \"4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56\" (UID: \"6c967b8a-ec86-413b-a25e-b81f3e9ac2dc\") " pod="openstack-operators/4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56" Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.439528 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2q9f\" (UniqueName: \"kubernetes.io/projected/6c967b8a-ec86-413b-a25e-b81f3e9ac2dc-kube-api-access-r2q9f\") pod \"4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56\" (UID: \"6c967b8a-ec86-413b-a25e-b81f3e9ac2dc\") " pod="openstack-operators/4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56" Oct 13 08:57:45 crc kubenswrapper[4685]: I1013 08:57:45.559268 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56" Oct 13 08:57:46 crc kubenswrapper[4685]: I1013 08:57:46.011767 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56"] Oct 13 08:57:46 crc kubenswrapper[4685]: I1013 08:57:46.270644 4685 generic.go:334] "Generic (PLEG): container finished" podID="6c967b8a-ec86-413b-a25e-b81f3e9ac2dc" containerID="193c4d4bd925c1dbeb464b87154c69e807166d1761e4c742834836f9ec3c2c76" exitCode=0 Oct 13 08:57:46 crc kubenswrapper[4685]: I1013 08:57:46.270733 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56" event={"ID":"6c967b8a-ec86-413b-a25e-b81f3e9ac2dc","Type":"ContainerDied","Data":"193c4d4bd925c1dbeb464b87154c69e807166d1761e4c742834836f9ec3c2c76"} Oct 13 08:57:46 crc kubenswrapper[4685]: I1013 08:57:46.270781 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56" event={"ID":"6c967b8a-ec86-413b-a25e-b81f3e9ac2dc","Type":"ContainerStarted","Data":"fdefb77d45f993df8374d7f05fb9bdd404f0b9ab498427503329c7335a23ccbc"} Oct 13 08:57:47 crc kubenswrapper[4685]: I1013 08:57:47.278930 4685 generic.go:334] "Generic (PLEG): container finished" podID="6c967b8a-ec86-413b-a25e-b81f3e9ac2dc" containerID="d9eb845d3d7a8c780786c26346a0aa1a9205484ab131233246c6a86eb9e609ac" exitCode=0 Oct 13 08:57:47 crc kubenswrapper[4685]: I1013 08:57:47.278970 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56" event={"ID":"6c967b8a-ec86-413b-a25e-b81f3e9ac2dc","Type":"ContainerDied","Data":"d9eb845d3d7a8c780786c26346a0aa1a9205484ab131233246c6a86eb9e609ac"} Oct 13 08:57:48 crc kubenswrapper[4685]: I1013 08:57:48.288122 4685 generic.go:334] "Generic (PLEG): container finished" podID="6c967b8a-ec86-413b-a25e-b81f3e9ac2dc" containerID="dbb9091264a4c3f48d048d75263b4a0bc5769161c01c705726f57a27c1f113e6" exitCode=0 Oct 13 08:57:48 crc kubenswrapper[4685]: I1013 08:57:48.288195 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56" event={"ID":"6c967b8a-ec86-413b-a25e-b81f3e9ac2dc","Type":"ContainerDied","Data":"dbb9091264a4c3f48d048d75263b4a0bc5769161c01c705726f57a27c1f113e6"} Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.380026 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9qj7h"] Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.381647 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9qj7h" Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.389861 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9qj7h"] Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.474281 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qj4d\" (UniqueName: \"kubernetes.io/projected/c6a9ded4-5eca-433e-acf8-679998ed41c3-kube-api-access-7qj4d\") pod \"redhat-operators-9qj7h\" (UID: \"c6a9ded4-5eca-433e-acf8-679998ed41c3\") " pod="openshift-marketplace/redhat-operators-9qj7h" Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.474337 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6a9ded4-5eca-433e-acf8-679998ed41c3-catalog-content\") pod \"redhat-operators-9qj7h\" (UID: \"c6a9ded4-5eca-433e-acf8-679998ed41c3\") " pod="openshift-marketplace/redhat-operators-9qj7h" Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.474418 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6a9ded4-5eca-433e-acf8-679998ed41c3-utilities\") pod \"redhat-operators-9qj7h\" (UID: \"c6a9ded4-5eca-433e-acf8-679998ed41c3\") " pod="openshift-marketplace/redhat-operators-9qj7h" Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.577735 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6a9ded4-5eca-433e-acf8-679998ed41c3-utilities\") pod \"redhat-operators-9qj7h\" (UID: \"c6a9ded4-5eca-433e-acf8-679998ed41c3\") " pod="openshift-marketplace/redhat-operators-9qj7h" Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.578306 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qj4d\" (UniqueName: \"kubernetes.io/projected/c6a9ded4-5eca-433e-acf8-679998ed41c3-kube-api-access-7qj4d\") pod \"redhat-operators-9qj7h\" (UID: \"c6a9ded4-5eca-433e-acf8-679998ed41c3\") " pod="openshift-marketplace/redhat-operators-9qj7h" Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.578347 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6a9ded4-5eca-433e-acf8-679998ed41c3-catalog-content\") pod \"redhat-operators-9qj7h\" (UID: \"c6a9ded4-5eca-433e-acf8-679998ed41c3\") " pod="openshift-marketplace/redhat-operators-9qj7h" Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.579229 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6a9ded4-5eca-433e-acf8-679998ed41c3-utilities\") pod \"redhat-operators-9qj7h\" (UID: \"c6a9ded4-5eca-433e-acf8-679998ed41c3\") " pod="openshift-marketplace/redhat-operators-9qj7h" Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.579583 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6a9ded4-5eca-433e-acf8-679998ed41c3-catalog-content\") pod \"redhat-operators-9qj7h\" (UID: \"c6a9ded4-5eca-433e-acf8-679998ed41c3\") " pod="openshift-marketplace/redhat-operators-9qj7h" Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.600776 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qj4d\" (UniqueName: \"kubernetes.io/projected/c6a9ded4-5eca-433e-acf8-679998ed41c3-kube-api-access-7qj4d\") pod \"redhat-operators-9qj7h\" (UID: \"c6a9ded4-5eca-433e-acf8-679998ed41c3\") " pod="openshift-marketplace/redhat-operators-9qj7h" Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.631530 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56" Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.679221 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c967b8a-ec86-413b-a25e-b81f3e9ac2dc-util\") pod \"6c967b8a-ec86-413b-a25e-b81f3e9ac2dc\" (UID: \"6c967b8a-ec86-413b-a25e-b81f3e9ac2dc\") " Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.679419 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2q9f\" (UniqueName: \"kubernetes.io/projected/6c967b8a-ec86-413b-a25e-b81f3e9ac2dc-kube-api-access-r2q9f\") pod \"6c967b8a-ec86-413b-a25e-b81f3e9ac2dc\" (UID: \"6c967b8a-ec86-413b-a25e-b81f3e9ac2dc\") " Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.679480 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c967b8a-ec86-413b-a25e-b81f3e9ac2dc-bundle\") pod \"6c967b8a-ec86-413b-a25e-b81f3e9ac2dc\" (UID: \"6c967b8a-ec86-413b-a25e-b81f3e9ac2dc\") " Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.680312 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c967b8a-ec86-413b-a25e-b81f3e9ac2dc-bundle" (OuterVolumeSpecName: "bundle") pod "6c967b8a-ec86-413b-a25e-b81f3e9ac2dc" (UID: "6c967b8a-ec86-413b-a25e-b81f3e9ac2dc"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.694127 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c967b8a-ec86-413b-a25e-b81f3e9ac2dc-util" (OuterVolumeSpecName: "util") pod "6c967b8a-ec86-413b-a25e-b81f3e9ac2dc" (UID: "6c967b8a-ec86-413b-a25e-b81f3e9ac2dc"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.695704 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c967b8a-ec86-413b-a25e-b81f3e9ac2dc-kube-api-access-r2q9f" (OuterVolumeSpecName: "kube-api-access-r2q9f") pod "6c967b8a-ec86-413b-a25e-b81f3e9ac2dc" (UID: "6c967b8a-ec86-413b-a25e-b81f3e9ac2dc"). InnerVolumeSpecName "kube-api-access-r2q9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.712769 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9qj7h" Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.781111 4685 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c967b8a-ec86-413b-a25e-b81f3e9ac2dc-util\") on node \"crc\" DevicePath \"\"" Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.781142 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2q9f\" (UniqueName: \"kubernetes.io/projected/6c967b8a-ec86-413b-a25e-b81f3e9ac2dc-kube-api-access-r2q9f\") on node \"crc\" DevicePath \"\"" Oct 13 08:57:49 crc kubenswrapper[4685]: I1013 08:57:49.781153 4685 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c967b8a-ec86-413b-a25e-b81f3e9ac2dc-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 08:57:50 crc kubenswrapper[4685]: I1013 08:57:50.137501 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9qj7h"] Oct 13 08:57:50 crc kubenswrapper[4685]: W1013 08:57:50.145193 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6a9ded4_5eca_433e_acf8_679998ed41c3.slice/crio-59885a246ffa6c2746effea8098e64819c6d1d6d5adab89b3fc41ace3556dd79 WatchSource:0}: Error finding container 59885a246ffa6c2746effea8098e64819c6d1d6d5adab89b3fc41ace3556dd79: Status 404 returned error can't find the container with id 59885a246ffa6c2746effea8098e64819c6d1d6d5adab89b3fc41ace3556dd79 Oct 13 08:57:50 crc kubenswrapper[4685]: I1013 08:57:50.165514 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z9vp4"] Oct 13 08:57:50 crc kubenswrapper[4685]: I1013 08:57:50.165721 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-z9vp4" podUID="0141650e-9da4-48d3-a6d8-d0d238397f97" containerName="registry-server" containerID="cri-o://47d5d87689bea2586fa6f54967b4a9e7115593808d1ee38bad3dab66934544f3" gracePeriod=2 Oct 13 08:57:50 crc kubenswrapper[4685]: I1013 08:57:50.299523 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56" event={"ID":"6c967b8a-ec86-413b-a25e-b81f3e9ac2dc","Type":"ContainerDied","Data":"fdefb77d45f993df8374d7f05fb9bdd404f0b9ab498427503329c7335a23ccbc"} Oct 13 08:57:50 crc kubenswrapper[4685]: I1013 08:57:50.299558 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdefb77d45f993df8374d7f05fb9bdd404f0b9ab498427503329c7335a23ccbc" Oct 13 08:57:50 crc kubenswrapper[4685]: I1013 08:57:50.299563 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56" Oct 13 08:57:50 crc kubenswrapper[4685]: I1013 08:57:50.300444 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qj7h" event={"ID":"c6a9ded4-5eca-433e-acf8-679998ed41c3","Type":"ContainerStarted","Data":"59885a246ffa6c2746effea8098e64819c6d1d6d5adab89b3fc41ace3556dd79"} Oct 13 08:57:50 crc kubenswrapper[4685]: I1013 08:57:50.324932 4685 generic.go:334] "Generic (PLEG): container finished" podID="0141650e-9da4-48d3-a6d8-d0d238397f97" containerID="47d5d87689bea2586fa6f54967b4a9e7115593808d1ee38bad3dab66934544f3" exitCode=0 Oct 13 08:57:50 crc kubenswrapper[4685]: I1013 08:57:50.324976 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z9vp4" event={"ID":"0141650e-9da4-48d3-a6d8-d0d238397f97","Type":"ContainerDied","Data":"47d5d87689bea2586fa6f54967b4a9e7115593808d1ee38bad3dab66934544f3"} Oct 13 08:57:50 crc kubenswrapper[4685]: I1013 08:57:50.518993 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z9vp4" Oct 13 08:57:50 crc kubenswrapper[4685]: I1013 08:57:50.590055 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0141650e-9da4-48d3-a6d8-d0d238397f97-catalog-content\") pod \"0141650e-9da4-48d3-a6d8-d0d238397f97\" (UID: \"0141650e-9da4-48d3-a6d8-d0d238397f97\") " Oct 13 08:57:50 crc kubenswrapper[4685]: I1013 08:57:50.590116 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0141650e-9da4-48d3-a6d8-d0d238397f97-utilities\") pod \"0141650e-9da4-48d3-a6d8-d0d238397f97\" (UID: \"0141650e-9da4-48d3-a6d8-d0d238397f97\") " Oct 13 08:57:50 crc kubenswrapper[4685]: I1013 08:57:50.590163 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngkfz\" (UniqueName: \"kubernetes.io/projected/0141650e-9da4-48d3-a6d8-d0d238397f97-kube-api-access-ngkfz\") pod \"0141650e-9da4-48d3-a6d8-d0d238397f97\" (UID: \"0141650e-9da4-48d3-a6d8-d0d238397f97\") " Oct 13 08:57:50 crc kubenswrapper[4685]: I1013 08:57:50.591259 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0141650e-9da4-48d3-a6d8-d0d238397f97-utilities" (OuterVolumeSpecName: "utilities") pod "0141650e-9da4-48d3-a6d8-d0d238397f97" (UID: "0141650e-9da4-48d3-a6d8-d0d238397f97"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:57:50 crc kubenswrapper[4685]: I1013 08:57:50.595561 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0141650e-9da4-48d3-a6d8-d0d238397f97-kube-api-access-ngkfz" (OuterVolumeSpecName: "kube-api-access-ngkfz") pod "0141650e-9da4-48d3-a6d8-d0d238397f97" (UID: "0141650e-9da4-48d3-a6d8-d0d238397f97"). InnerVolumeSpecName "kube-api-access-ngkfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:57:50 crc kubenswrapper[4685]: I1013 08:57:50.601982 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0141650e-9da4-48d3-a6d8-d0d238397f97-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0141650e-9da4-48d3-a6d8-d0d238397f97" (UID: "0141650e-9da4-48d3-a6d8-d0d238397f97"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:57:50 crc kubenswrapper[4685]: I1013 08:57:50.691272 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0141650e-9da4-48d3-a6d8-d0d238397f97-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 08:57:50 crc kubenswrapper[4685]: I1013 08:57:50.691299 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0141650e-9da4-48d3-a6d8-d0d238397f97-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 08:57:50 crc kubenswrapper[4685]: I1013 08:57:50.691308 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngkfz\" (UniqueName: \"kubernetes.io/projected/0141650e-9da4-48d3-a6d8-d0d238397f97-kube-api-access-ngkfz\") on node \"crc\" DevicePath \"\"" Oct 13 08:57:51 crc kubenswrapper[4685]: I1013 08:57:51.335446 4685 generic.go:334] "Generic (PLEG): container finished" podID="c6a9ded4-5eca-433e-acf8-679998ed41c3" containerID="fb987bc5ecd151f7381f57583438f44aff1378406a8bfaa0ee8ec2b35ca7b464" exitCode=0 Oct 13 08:57:51 crc kubenswrapper[4685]: I1013 08:57:51.335593 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qj7h" event={"ID":"c6a9ded4-5eca-433e-acf8-679998ed41c3","Type":"ContainerDied","Data":"fb987bc5ecd151f7381f57583438f44aff1378406a8bfaa0ee8ec2b35ca7b464"} Oct 13 08:57:51 crc kubenswrapper[4685]: I1013 08:57:51.348085 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z9vp4" event={"ID":"0141650e-9da4-48d3-a6d8-d0d238397f97","Type":"ContainerDied","Data":"f86165a416e35b0c41ff049d9f631942c0c6ba02550f040ef4e2035122fa4ab4"} Oct 13 08:57:51 crc kubenswrapper[4685]: I1013 08:57:51.348376 4685 scope.go:117] "RemoveContainer" containerID="47d5d87689bea2586fa6f54967b4a9e7115593808d1ee38bad3dab66934544f3" Oct 13 08:57:51 crc kubenswrapper[4685]: I1013 08:57:51.348184 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z9vp4" Oct 13 08:57:51 crc kubenswrapper[4685]: I1013 08:57:51.393152 4685 scope.go:117] "RemoveContainer" containerID="3e13a94a7f5d614150cc45cd511c77e8f4f2c6f1fd4003bed7423804f4f73b02" Oct 13 08:57:51 crc kubenswrapper[4685]: I1013 08:57:51.396262 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z9vp4"] Oct 13 08:57:51 crc kubenswrapper[4685]: I1013 08:57:51.400847 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-z9vp4"] Oct 13 08:57:51 crc kubenswrapper[4685]: I1013 08:57:51.415791 4685 scope.go:117] "RemoveContainer" containerID="ba32490fa78c0658c7c8915bd33348dd97becd0d2a423a86742e7e8ee2431145" Oct 13 08:57:51 crc kubenswrapper[4685]: I1013 08:57:51.512559 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0141650e-9da4-48d3-a6d8-d0d238397f97" path="/var/lib/kubelet/pods/0141650e-9da4-48d3-a6d8-d0d238397f97/volumes" Oct 13 08:57:52 crc kubenswrapper[4685]: I1013 08:57:52.357485 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qj7h" event={"ID":"c6a9ded4-5eca-433e-acf8-679998ed41c3","Type":"ContainerStarted","Data":"f9d59933ca65187efe96a133f2f5912250a80afc17e627e25f76f645c749d509"} Oct 13 08:57:52 crc kubenswrapper[4685]: I1013 08:57:52.980691 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 08:57:52 crc kubenswrapper[4685]: I1013 08:57:52.980780 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 08:57:53 crc kubenswrapper[4685]: I1013 08:57:53.367906 4685 generic.go:334] "Generic (PLEG): container finished" podID="c6a9ded4-5eca-433e-acf8-679998ed41c3" containerID="f9d59933ca65187efe96a133f2f5912250a80afc17e627e25f76f645c749d509" exitCode=0 Oct 13 08:57:53 crc kubenswrapper[4685]: I1013 08:57:53.367974 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qj7h" event={"ID":"c6a9ded4-5eca-433e-acf8-679998ed41c3","Type":"ContainerDied","Data":"f9d59933ca65187efe96a133f2f5912250a80afc17e627e25f76f645c749d509"} Oct 13 08:57:54 crc kubenswrapper[4685]: I1013 08:57:54.374649 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qj7h" event={"ID":"c6a9ded4-5eca-433e-acf8-679998ed41c3","Type":"ContainerStarted","Data":"bf5c4ac9e9a31a7e0a77a0508215f812ef5d49d3c4f52aa73b515e92ef093f66"} Oct 13 08:57:54 crc kubenswrapper[4685]: I1013 08:57:54.392113 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9qj7h" podStartSLOduration=2.828111577 podStartE2EDuration="5.3920937s" podCreationTimestamp="2025-10-13 08:57:49 +0000 UTC" firstStartedPulling="2025-10-13 08:57:51.338311587 +0000 UTC m=+796.486187378" lastFinishedPulling="2025-10-13 08:57:53.90229374 +0000 UTC m=+799.050169501" observedRunningTime="2025-10-13 08:57:54.388207733 +0000 UTC m=+799.536083514" watchObservedRunningTime="2025-10-13 08:57:54.3920937 +0000 UTC m=+799.539969461" Oct 13 08:57:57 crc kubenswrapper[4685]: I1013 08:57:57.648709 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6c8cf6687-nbccj"] Oct 13 08:57:57 crc kubenswrapper[4685]: E1013 08:57:57.650412 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c967b8a-ec86-413b-a25e-b81f3e9ac2dc" containerName="pull" Oct 13 08:57:57 crc kubenswrapper[4685]: I1013 08:57:57.650511 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c967b8a-ec86-413b-a25e-b81f3e9ac2dc" containerName="pull" Oct 13 08:57:57 crc kubenswrapper[4685]: E1013 08:57:57.650589 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c967b8a-ec86-413b-a25e-b81f3e9ac2dc" containerName="util" Oct 13 08:57:57 crc kubenswrapper[4685]: I1013 08:57:57.650661 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c967b8a-ec86-413b-a25e-b81f3e9ac2dc" containerName="util" Oct 13 08:57:57 crc kubenswrapper[4685]: E1013 08:57:57.650735 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c967b8a-ec86-413b-a25e-b81f3e9ac2dc" containerName="extract" Oct 13 08:57:57 crc kubenswrapper[4685]: I1013 08:57:57.650831 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c967b8a-ec86-413b-a25e-b81f3e9ac2dc" containerName="extract" Oct 13 08:57:57 crc kubenswrapper[4685]: E1013 08:57:57.650930 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0141650e-9da4-48d3-a6d8-d0d238397f97" containerName="registry-server" Oct 13 08:57:57 crc kubenswrapper[4685]: I1013 08:57:57.651007 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="0141650e-9da4-48d3-a6d8-d0d238397f97" containerName="registry-server" Oct 13 08:57:57 crc kubenswrapper[4685]: E1013 08:57:57.651083 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0141650e-9da4-48d3-a6d8-d0d238397f97" containerName="extract-content" Oct 13 08:57:57 crc kubenswrapper[4685]: I1013 08:57:57.651157 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="0141650e-9da4-48d3-a6d8-d0d238397f97" containerName="extract-content" Oct 13 08:57:57 crc kubenswrapper[4685]: E1013 08:57:57.651230 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0141650e-9da4-48d3-a6d8-d0d238397f97" containerName="extract-utilities" Oct 13 08:57:57 crc kubenswrapper[4685]: I1013 08:57:57.651325 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="0141650e-9da4-48d3-a6d8-d0d238397f97" containerName="extract-utilities" Oct 13 08:57:57 crc kubenswrapper[4685]: I1013 08:57:57.651514 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c967b8a-ec86-413b-a25e-b81f3e9ac2dc" containerName="extract" Oct 13 08:57:57 crc kubenswrapper[4685]: I1013 08:57:57.651607 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="0141650e-9da4-48d3-a6d8-d0d238397f97" containerName="registry-server" Oct 13 08:57:57 crc kubenswrapper[4685]: I1013 08:57:57.652414 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6c8cf6687-nbccj" Oct 13 08:57:57 crc kubenswrapper[4685]: I1013 08:57:57.655141 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-gclxr" Oct 13 08:57:57 crc kubenswrapper[4685]: I1013 08:57:57.685849 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6c8cf6687-nbccj"] Oct 13 08:57:57 crc kubenswrapper[4685]: I1013 08:57:57.773606 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtcsg\" (UniqueName: \"kubernetes.io/projected/c31c40db-9982-4692-91cc-26bf7b4ba509-kube-api-access-gtcsg\") pod \"openstack-operator-controller-operator-6c8cf6687-nbccj\" (UID: \"c31c40db-9982-4692-91cc-26bf7b4ba509\") " pod="openstack-operators/openstack-operator-controller-operator-6c8cf6687-nbccj" Oct 13 08:57:57 crc kubenswrapper[4685]: I1013 08:57:57.874362 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtcsg\" (UniqueName: \"kubernetes.io/projected/c31c40db-9982-4692-91cc-26bf7b4ba509-kube-api-access-gtcsg\") pod \"openstack-operator-controller-operator-6c8cf6687-nbccj\" (UID: \"c31c40db-9982-4692-91cc-26bf7b4ba509\") " pod="openstack-operators/openstack-operator-controller-operator-6c8cf6687-nbccj" Oct 13 08:57:57 crc kubenswrapper[4685]: I1013 08:57:57.892520 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtcsg\" (UniqueName: \"kubernetes.io/projected/c31c40db-9982-4692-91cc-26bf7b4ba509-kube-api-access-gtcsg\") pod \"openstack-operator-controller-operator-6c8cf6687-nbccj\" (UID: \"c31c40db-9982-4692-91cc-26bf7b4ba509\") " pod="openstack-operators/openstack-operator-controller-operator-6c8cf6687-nbccj" Oct 13 08:57:57 crc kubenswrapper[4685]: I1013 08:57:57.970679 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6c8cf6687-nbccj" Oct 13 08:57:58 crc kubenswrapper[4685]: I1013 08:57:58.527995 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6c8cf6687-nbccj"] Oct 13 08:57:59 crc kubenswrapper[4685]: I1013 08:57:59.415628 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6c8cf6687-nbccj" event={"ID":"c31c40db-9982-4692-91cc-26bf7b4ba509","Type":"ContainerStarted","Data":"897731972b53e9da25857b89ea0c6c86d92640bc9eb81e3eb75c9513c26a03e1"} Oct 13 08:57:59 crc kubenswrapper[4685]: I1013 08:57:59.713383 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9qj7h" Oct 13 08:57:59 crc kubenswrapper[4685]: I1013 08:57:59.713508 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9qj7h" Oct 13 08:57:59 crc kubenswrapper[4685]: I1013 08:57:59.757116 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9qj7h" Oct 13 08:58:00 crc kubenswrapper[4685]: I1013 08:58:00.476518 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9qj7h" Oct 13 08:58:02 crc kubenswrapper[4685]: I1013 08:58:02.172130 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9qj7h"] Oct 13 08:58:03 crc kubenswrapper[4685]: I1013 08:58:03.442575 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6c8cf6687-nbccj" event={"ID":"c31c40db-9982-4692-91cc-26bf7b4ba509","Type":"ContainerStarted","Data":"50db3a3442e38eff030145975273b0427f813be95b9db6b0ab8446f2c69ca6b4"} Oct 13 08:58:03 crc kubenswrapper[4685]: I1013 08:58:03.442657 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9qj7h" podUID="c6a9ded4-5eca-433e-acf8-679998ed41c3" containerName="registry-server" containerID="cri-o://bf5c4ac9e9a31a7e0a77a0508215f812ef5d49d3c4f52aa73b515e92ef093f66" gracePeriod=2 Oct 13 08:58:04 crc kubenswrapper[4685]: I1013 08:58:04.451056 4685 generic.go:334] "Generic (PLEG): container finished" podID="c6a9ded4-5eca-433e-acf8-679998ed41c3" containerID="bf5c4ac9e9a31a7e0a77a0508215f812ef5d49d3c4f52aa73b515e92ef093f66" exitCode=0 Oct 13 08:58:04 crc kubenswrapper[4685]: I1013 08:58:04.451098 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qj7h" event={"ID":"c6a9ded4-5eca-433e-acf8-679998ed41c3","Type":"ContainerDied","Data":"bf5c4ac9e9a31a7e0a77a0508215f812ef5d49d3c4f52aa73b515e92ef093f66"} Oct 13 08:58:05 crc kubenswrapper[4685]: I1013 08:58:05.099154 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9qj7h" Oct 13 08:58:05 crc kubenswrapper[4685]: I1013 08:58:05.281660 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6a9ded4-5eca-433e-acf8-679998ed41c3-catalog-content\") pod \"c6a9ded4-5eca-433e-acf8-679998ed41c3\" (UID: \"c6a9ded4-5eca-433e-acf8-679998ed41c3\") " Oct 13 08:58:05 crc kubenswrapper[4685]: I1013 08:58:05.282001 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qj4d\" (UniqueName: \"kubernetes.io/projected/c6a9ded4-5eca-433e-acf8-679998ed41c3-kube-api-access-7qj4d\") pod \"c6a9ded4-5eca-433e-acf8-679998ed41c3\" (UID: \"c6a9ded4-5eca-433e-acf8-679998ed41c3\") " Oct 13 08:58:05 crc kubenswrapper[4685]: I1013 08:58:05.282058 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6a9ded4-5eca-433e-acf8-679998ed41c3-utilities\") pod \"c6a9ded4-5eca-433e-acf8-679998ed41c3\" (UID: \"c6a9ded4-5eca-433e-acf8-679998ed41c3\") " Oct 13 08:58:05 crc kubenswrapper[4685]: I1013 08:58:05.283840 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6a9ded4-5eca-433e-acf8-679998ed41c3-utilities" (OuterVolumeSpecName: "utilities") pod "c6a9ded4-5eca-433e-acf8-679998ed41c3" (UID: "c6a9ded4-5eca-433e-acf8-679998ed41c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:58:05 crc kubenswrapper[4685]: I1013 08:58:05.289426 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6a9ded4-5eca-433e-acf8-679998ed41c3-kube-api-access-7qj4d" (OuterVolumeSpecName: "kube-api-access-7qj4d") pod "c6a9ded4-5eca-433e-acf8-679998ed41c3" (UID: "c6a9ded4-5eca-433e-acf8-679998ed41c3"). InnerVolumeSpecName "kube-api-access-7qj4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 08:58:05 crc kubenswrapper[4685]: I1013 08:58:05.384759 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qj4d\" (UniqueName: \"kubernetes.io/projected/c6a9ded4-5eca-433e-acf8-679998ed41c3-kube-api-access-7qj4d\") on node \"crc\" DevicePath \"\"" Oct 13 08:58:05 crc kubenswrapper[4685]: I1013 08:58:05.384843 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6a9ded4-5eca-433e-acf8-679998ed41c3-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 08:58:05 crc kubenswrapper[4685]: I1013 08:58:05.403087 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6a9ded4-5eca-433e-acf8-679998ed41c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c6a9ded4-5eca-433e-acf8-679998ed41c3" (UID: "c6a9ded4-5eca-433e-acf8-679998ed41c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 08:58:05 crc kubenswrapper[4685]: I1013 08:58:05.458161 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qj7h" event={"ID":"c6a9ded4-5eca-433e-acf8-679998ed41c3","Type":"ContainerDied","Data":"59885a246ffa6c2746effea8098e64819c6d1d6d5adab89b3fc41ace3556dd79"} Oct 13 08:58:05 crc kubenswrapper[4685]: I1013 08:58:05.458221 4685 scope.go:117] "RemoveContainer" containerID="bf5c4ac9e9a31a7e0a77a0508215f812ef5d49d3c4f52aa73b515e92ef093f66" Oct 13 08:58:05 crc kubenswrapper[4685]: I1013 08:58:05.458281 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9qj7h" Oct 13 08:58:05 crc kubenswrapper[4685]: I1013 08:58:05.488013 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6a9ded4-5eca-433e-acf8-679998ed41c3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 08:58:05 crc kubenswrapper[4685]: I1013 08:58:05.497147 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9qj7h"] Oct 13 08:58:05 crc kubenswrapper[4685]: I1013 08:58:05.500641 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9qj7h"] Oct 13 08:58:05 crc kubenswrapper[4685]: I1013 08:58:05.510626 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6a9ded4-5eca-433e-acf8-679998ed41c3" path="/var/lib/kubelet/pods/c6a9ded4-5eca-433e-acf8-679998ed41c3/volumes" Oct 13 08:58:06 crc kubenswrapper[4685]: I1013 08:58:06.259893 4685 scope.go:117] "RemoveContainer" containerID="f9d59933ca65187efe96a133f2f5912250a80afc17e627e25f76f645c749d509" Oct 13 08:58:06 crc kubenswrapper[4685]: I1013 08:58:06.691157 4685 scope.go:117] "RemoveContainer" containerID="fb987bc5ecd151f7381f57583438f44aff1378406a8bfaa0ee8ec2b35ca7b464" Oct 13 08:58:07 crc kubenswrapper[4685]: I1013 08:58:07.480411 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6c8cf6687-nbccj" event={"ID":"c31c40db-9982-4692-91cc-26bf7b4ba509","Type":"ContainerStarted","Data":"807d2c62ca4857f59a4900f515ccca0ea7a036d06f01230dde448152a82f9b5c"} Oct 13 08:58:07 crc kubenswrapper[4685]: I1013 08:58:07.480750 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6c8cf6687-nbccj" Oct 13 08:58:07 crc kubenswrapper[4685]: I1013 08:58:07.536092 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6c8cf6687-nbccj" podStartSLOduration=2.33273753 podStartE2EDuration="10.536073351s" podCreationTimestamp="2025-10-13 08:57:57 +0000 UTC" firstStartedPulling="2025-10-13 08:57:58.541534486 +0000 UTC m=+803.689410247" lastFinishedPulling="2025-10-13 08:58:06.744870307 +0000 UTC m=+811.892746068" observedRunningTime="2025-10-13 08:58:07.533052068 +0000 UTC m=+812.680927839" watchObservedRunningTime="2025-10-13 08:58:07.536073351 +0000 UTC m=+812.683949112" Oct 13 08:58:08 crc kubenswrapper[4685]: I1013 08:58:08.492334 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6c8cf6687-nbccj" Oct 13 08:58:22 crc kubenswrapper[4685]: I1013 08:58:22.980167 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 08:58:22 crc kubenswrapper[4685]: I1013 08:58:22.980863 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.813411 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-5d7zf"] Oct 13 08:58:42 crc kubenswrapper[4685]: E1013 08:58:42.814310 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6a9ded4-5eca-433e-acf8-679998ed41c3" containerName="extract-content" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.814327 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6a9ded4-5eca-433e-acf8-679998ed41c3" containerName="extract-content" Oct 13 08:58:42 crc kubenswrapper[4685]: E1013 08:58:42.814345 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6a9ded4-5eca-433e-acf8-679998ed41c3" containerName="registry-server" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.814353 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6a9ded4-5eca-433e-acf8-679998ed41c3" containerName="registry-server" Oct 13 08:58:42 crc kubenswrapper[4685]: E1013 08:58:42.814366 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6a9ded4-5eca-433e-acf8-679998ed41c3" containerName="extract-utilities" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.814375 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6a9ded4-5eca-433e-acf8-679998ed41c3" containerName="extract-utilities" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.814506 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6a9ded4-5eca-433e-acf8-679998ed41c3" containerName="registry-server" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.815283 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-5d7zf" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.818373 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-qbl58"] Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.819595 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-qbl58" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.822327 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-l7jp6" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.824188 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-lprkc" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.884761 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-5d7zf"] Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.907954 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-zm9bd"] Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.908809 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zm9bd" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.909563 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-hbnzh"] Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.910659 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfj4l\" (UniqueName: \"kubernetes.io/projected/54405c7c-61f7-41ea-ae0a-29128b51326c-kube-api-access-xfj4l\") pod \"glance-operator-controller-manager-7bb46cd7d-zm9bd\" (UID: \"54405c7c-61f7-41ea-ae0a-29128b51326c\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zm9bd" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.910703 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54vlr\" (UniqueName: \"kubernetes.io/projected/c92c1fba-b02b-4b6c-9570-b75ee60c5e86-kube-api-access-54vlr\") pod \"cinder-operator-controller-manager-59cdc64769-qbl58\" (UID: \"c92c1fba-b02b-4b6c-9570-b75ee60c5e86\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-qbl58" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.910798 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwkjl\" (UniqueName: \"kubernetes.io/projected/e509e801-67de-4a55-bd22-cf3f73deca81-kube-api-access-wwkjl\") pod \"barbican-operator-controller-manager-64f84fcdbb-5d7zf\" (UID: \"e509e801-67de-4a55-bd22-cf3f73deca81\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-5d7zf" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.910807 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-hbnzh" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.914132 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-mj7x8"] Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.915258 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-mj7x8" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.920237 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-p8h48" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.920649 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-tpfnw" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.920772 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-bmc5v" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.922887 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-sl5lv"] Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.923732 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-sl5lv" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.926932 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-hbnzh"] Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.927145 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-4xk7p" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.938455 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-mj7x8"] Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.968399 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-qbl58"] Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.978363 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-2bcss"] Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.979228 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-2bcss" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.981154 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-5wpkm" Oct 13 08:58:42 crc kubenswrapper[4685]: I1013 08:58:42.983577 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.015517 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwkjl\" (UniqueName: \"kubernetes.io/projected/e509e801-67de-4a55-bd22-cf3f73deca81-kube-api-access-wwkjl\") pod \"barbican-operator-controller-manager-64f84fcdbb-5d7zf\" (UID: \"e509e801-67de-4a55-bd22-cf3f73deca81\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-5d7zf" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.015570 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfj4l\" (UniqueName: \"kubernetes.io/projected/54405c7c-61f7-41ea-ae0a-29128b51326c-kube-api-access-xfj4l\") pod \"glance-operator-controller-manager-7bb46cd7d-zm9bd\" (UID: \"54405c7c-61f7-41ea-ae0a-29128b51326c\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zm9bd" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.015595 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54vlr\" (UniqueName: \"kubernetes.io/projected/c92c1fba-b02b-4b6c-9570-b75ee60c5e86-kube-api-access-54vlr\") pod \"cinder-operator-controller-manager-59cdc64769-qbl58\" (UID: \"c92c1fba-b02b-4b6c-9570-b75ee60c5e86\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-qbl58" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.016040 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-xf9rd"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.017029 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-xf9rd" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.021271 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-wlvvz" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.027199 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-zm9bd"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.031084 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-sl5lv"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.042988 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-2bcss"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.061015 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.062191 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.067822 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54vlr\" (UniqueName: \"kubernetes.io/projected/c92c1fba-b02b-4b6c-9570-b75ee60c5e86-kube-api-access-54vlr\") pod \"cinder-operator-controller-manager-59cdc64769-qbl58\" (UID: \"c92c1fba-b02b-4b6c-9570-b75ee60c5e86\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-qbl58" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.068180 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-s7dfd" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.069003 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfj4l\" (UniqueName: \"kubernetes.io/projected/54405c7c-61f7-41ea-ae0a-29128b51326c-kube-api-access-xfj4l\") pod \"glance-operator-controller-manager-7bb46cd7d-zm9bd\" (UID: \"54405c7c-61f7-41ea-ae0a-29128b51326c\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zm9bd" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.070173 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-q8dkt"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.071134 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-59578bc799-q8dkt" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.073650 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwkjl\" (UniqueName: \"kubernetes.io/projected/e509e801-67de-4a55-bd22-cf3f73deca81-kube-api-access-wwkjl\") pod \"barbican-operator-controller-manager-64f84fcdbb-5d7zf\" (UID: \"e509e801-67de-4a55-bd22-cf3f73deca81\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-5d7zf" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.090826 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-2nnv4" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.117389 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7nx6\" (UniqueName: \"kubernetes.io/projected/a3e83b7f-5017-44e7-b507-f46a2d3f5488-kube-api-access-d7nx6\") pod \"infra-operator-controller-manager-585fc5b659-2bcss\" (UID: \"a3e83b7f-5017-44e7-b507-f46a2d3f5488\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-2bcss" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.117453 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a3e83b7f-5017-44e7-b507-f46a2d3f5488-cert\") pod \"infra-operator-controller-manager-585fc5b659-2bcss\" (UID: \"a3e83b7f-5017-44e7-b507-f46a2d3f5488\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-2bcss" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.117493 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7dbl\" (UniqueName: \"kubernetes.io/projected/eb08316b-7fd3-4d65-88e1-bbc91efcb7c7-kube-api-access-q7dbl\") pod \"horizon-operator-controller-manager-6d74794d9b-sl5lv\" (UID: \"eb08316b-7fd3-4d65-88e1-bbc91efcb7c7\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-sl5lv" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.117546 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hxjg\" (UniqueName: \"kubernetes.io/projected/92a042d7-669e-48d2-8d7e-8a8da4fc01eb-kube-api-access-6hxjg\") pod \"designate-operator-controller-manager-687df44cdb-hbnzh\" (UID: \"92a042d7-669e-48d2-8d7e-8a8da4fc01eb\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-hbnzh" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.117616 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqdmr\" (UniqueName: \"kubernetes.io/projected/6b61f732-3f39-459e-bec0-a6bdd5added4-kube-api-access-lqdmr\") pod \"heat-operator-controller-manager-6d9967f8dd-mj7x8\" (UID: \"6b61f732-3f39-459e-bec0-a6bdd5added4\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-mj7x8" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.138482 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-5d7zf" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.144722 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-qbl58" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.151163 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.175673 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-q8dkt"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.204299 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-xf9rd"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.219943 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqdmr\" (UniqueName: \"kubernetes.io/projected/6b61f732-3f39-459e-bec0-a6bdd5added4-kube-api-access-lqdmr\") pod \"heat-operator-controller-manager-6d9967f8dd-mj7x8\" (UID: \"6b61f732-3f39-459e-bec0-a6bdd5added4\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-mj7x8" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.220016 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqbtq\" (UniqueName: \"kubernetes.io/projected/b001d17a-1aea-44ba-86c5-ba6b312156c1-kube-api-access-cqbtq\") pod \"keystone-operator-controller-manager-f9d897d75-xtnqp\" (UID: \"b001d17a-1aea-44ba-86c5-ba6b312156c1\") " pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.220046 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7nx6\" (UniqueName: \"kubernetes.io/projected/a3e83b7f-5017-44e7-b507-f46a2d3f5488-kube-api-access-d7nx6\") pod \"infra-operator-controller-manager-585fc5b659-2bcss\" (UID: \"a3e83b7f-5017-44e7-b507-f46a2d3f5488\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-2bcss" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.220448 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a3e83b7f-5017-44e7-b507-f46a2d3f5488-cert\") pod \"infra-operator-controller-manager-585fc5b659-2bcss\" (UID: \"a3e83b7f-5017-44e7-b507-f46a2d3f5488\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-2bcss" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.220564 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7dbl\" (UniqueName: \"kubernetes.io/projected/eb08316b-7fd3-4d65-88e1-bbc91efcb7c7-kube-api-access-q7dbl\") pod \"horizon-operator-controller-manager-6d74794d9b-sl5lv\" (UID: \"eb08316b-7fd3-4d65-88e1-bbc91efcb7c7\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-sl5lv" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.220599 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsszh\" (UniqueName: \"kubernetes.io/projected/9899b0d7-81a4-49f7-91cd-3c5aa72d49b2-kube-api-access-zsszh\") pod \"ironic-operator-controller-manager-74cb5cbc49-xf9rd\" (UID: \"9899b0d7-81a4-49f7-91cd-3c5aa72d49b2\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-xf9rd" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.220656 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q58p\" (UniqueName: \"kubernetes.io/projected/12dc02ce-76d4-4376-ab6f-b2a75580cc4d-kube-api-access-8q58p\") pod \"manila-operator-controller-manager-59578bc799-q8dkt\" (UID: \"12dc02ce-76d4-4376-ab6f-b2a75580cc4d\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-q8dkt" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.220676 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hxjg\" (UniqueName: \"kubernetes.io/projected/92a042d7-669e-48d2-8d7e-8a8da4fc01eb-kube-api-access-6hxjg\") pod \"designate-operator-controller-manager-687df44cdb-hbnzh\" (UID: \"92a042d7-669e-48d2-8d7e-8a8da4fc01eb\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-hbnzh" Oct 13 08:58:43 crc kubenswrapper[4685]: E1013 08:58:43.221024 4685 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 13 08:58:43 crc kubenswrapper[4685]: E1013 08:58:43.221097 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a3e83b7f-5017-44e7-b507-f46a2d3f5488-cert podName:a3e83b7f-5017-44e7-b507-f46a2d3f5488 nodeName:}" failed. No retries permitted until 2025-10-13 08:58:43.721077604 +0000 UTC m=+848.868953365 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a3e83b7f-5017-44e7-b507-f46a2d3f5488-cert") pod "infra-operator-controller-manager-585fc5b659-2bcss" (UID: "a3e83b7f-5017-44e7-b507-f46a2d3f5488") : secret "infra-operator-webhook-server-cert" not found Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.234442 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zm9bd" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.239419 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-j65jt"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.240538 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-j65jt" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.256503 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-6zwqj" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.260508 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqdmr\" (UniqueName: \"kubernetes.io/projected/6b61f732-3f39-459e-bec0-a6bdd5added4-kube-api-access-lqdmr\") pod \"heat-operator-controller-manager-6d9967f8dd-mj7x8\" (UID: \"6b61f732-3f39-459e-bec0-a6bdd5added4\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-mj7x8" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.272892 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-j65jt"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.279469 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7nx6\" (UniqueName: \"kubernetes.io/projected/a3e83b7f-5017-44e7-b507-f46a2d3f5488-kube-api-access-d7nx6\") pod \"infra-operator-controller-manager-585fc5b659-2bcss\" (UID: \"a3e83b7f-5017-44e7-b507-f46a2d3f5488\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-2bcss" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.282454 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hxjg\" (UniqueName: \"kubernetes.io/projected/92a042d7-669e-48d2-8d7e-8a8da4fc01eb-kube-api-access-6hxjg\") pod \"designate-operator-controller-manager-687df44cdb-hbnzh\" (UID: \"92a042d7-669e-48d2-8d7e-8a8da4fc01eb\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-hbnzh" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.289566 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7dbl\" (UniqueName: \"kubernetes.io/projected/eb08316b-7fd3-4d65-88e1-bbc91efcb7c7-kube-api-access-q7dbl\") pod \"horizon-operator-controller-manager-6d74794d9b-sl5lv\" (UID: \"eb08316b-7fd3-4d65-88e1-bbc91efcb7c7\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-sl5lv" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.318728 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-cbwvq"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.319804 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-cbwvq" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.322621 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqbtq\" (UniqueName: \"kubernetes.io/projected/b001d17a-1aea-44ba-86c5-ba6b312156c1-kube-api-access-cqbtq\") pod \"keystone-operator-controller-manager-f9d897d75-xtnqp\" (UID: \"b001d17a-1aea-44ba-86c5-ba6b312156c1\") " pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.322680 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsszh\" (UniqueName: \"kubernetes.io/projected/9899b0d7-81a4-49f7-91cd-3c5aa72d49b2-kube-api-access-zsszh\") pod \"ironic-operator-controller-manager-74cb5cbc49-xf9rd\" (UID: \"9899b0d7-81a4-49f7-91cd-3c5aa72d49b2\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-xf9rd" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.322703 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q58p\" (UniqueName: \"kubernetes.io/projected/12dc02ce-76d4-4376-ab6f-b2a75580cc4d-kube-api-access-8q58p\") pod \"manila-operator-controller-manager-59578bc799-q8dkt\" (UID: \"12dc02ce-76d4-4376-ab6f-b2a75580cc4d\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-q8dkt" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.323315 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-9z6vn" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.333370 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-ft77l"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.334565 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-ft77l" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.338375 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-8x775" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.354008 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-cbwvq"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.355428 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q58p\" (UniqueName: \"kubernetes.io/projected/12dc02ce-76d4-4376-ab6f-b2a75580cc4d-kube-api-access-8q58p\") pod \"manila-operator-controller-manager-59578bc799-q8dkt\" (UID: \"12dc02ce-76d4-4376-ab6f-b2a75580cc4d\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-q8dkt" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.359454 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqbtq\" (UniqueName: \"kubernetes.io/projected/b001d17a-1aea-44ba-86c5-ba6b312156c1-kube-api-access-cqbtq\") pod \"keystone-operator-controller-manager-f9d897d75-xtnqp\" (UID: \"b001d17a-1aea-44ba-86c5-ba6b312156c1\") " pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.369488 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsszh\" (UniqueName: \"kubernetes.io/projected/9899b0d7-81a4-49f7-91cd-3c5aa72d49b2-kube-api-access-zsszh\") pod \"ironic-operator-controller-manager-74cb5cbc49-xf9rd\" (UID: \"9899b0d7-81a4-49f7-91cd-3c5aa72d49b2\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-xf9rd" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.405971 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-b9jts"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.407206 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-b9jts" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.408307 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-ft77l"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.412278 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-sb8m4" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.423829 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8zpb\" (UniqueName: \"kubernetes.io/projected/de8c01ad-9e44-42f8-956c-0bb61165a222-kube-api-access-l8zpb\") pod \"neutron-operator-controller-manager-797d478b46-cbwvq\" (UID: \"de8c01ad-9e44-42f8-956c-0bb61165a222\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-cbwvq" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.429779 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktqc6\" (UniqueName: \"kubernetes.io/projected/e297e3a5-6cd9-4314-b4dc-76ba4c25c6f5-kube-api-access-ktqc6\") pod \"mariadb-operator-controller-manager-5777b4f897-j65jt\" (UID: \"e297e3a5-6cd9-4314-b4dc-76ba4c25c6f5\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-j65jt" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.433769 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.434808 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-b9jts"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.434901 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.441170 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.441407 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-g5ckf" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.441557 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.442114 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.459971 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-869cc7797f-7ktz6"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.461463 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-7ktz6" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.464902 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-56h7w" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.470449 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-59578bc799-q8dkt" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.480045 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-qnmxx"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.481311 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qnmxx" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.492433 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-d8gjz" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.534327 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72hqs\" (UniqueName: \"kubernetes.io/projected/d363a932-2a5d-4082-841a-b5d677d0b1a1-kube-api-access-72hqs\") pod \"nova-operator-controller-manager-57bb74c7bf-ft77l\" (UID: \"d363a932-2a5d-4082-841a-b5d677d0b1a1\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-ft77l" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.534366 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddkmn\" (UniqueName: \"kubernetes.io/projected/d4c8f24c-6e84-4931-8edb-504b184ea7b0-kube-api-access-ddkmn\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td\" (UID: \"d4c8f24c-6e84-4931-8edb-504b184ea7b0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.534473 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-282kw\" (UniqueName: \"kubernetes.io/projected/a4e11db1-f272-469d-9c22-a649cdbcf95e-kube-api-access-282kw\") pod \"ovn-operator-controller-manager-869cc7797f-7ktz6\" (UID: \"a4e11db1-f272-469d-9c22-a649cdbcf95e\") " pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-7ktz6" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.534488 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d4c8f24c-6e84-4931-8edb-504b184ea7b0-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td\" (UID: \"d4c8f24c-6e84-4931-8edb-504b184ea7b0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.534513 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nzvg\" (UniqueName: \"kubernetes.io/projected/f1085de4-f906-4315-b8b5-a3ee4e7182c1-kube-api-access-7nzvg\") pod \"telemetry-operator-controller-manager-578874c84d-qnmxx\" (UID: \"f1085de4-f906-4315-b8b5-a3ee4e7182c1\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qnmxx" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.534534 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8zpb\" (UniqueName: \"kubernetes.io/projected/de8c01ad-9e44-42f8-956c-0bb61165a222-kube-api-access-l8zpb\") pod \"neutron-operator-controller-manager-797d478b46-cbwvq\" (UID: \"de8c01ad-9e44-42f8-956c-0bb61165a222\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-cbwvq" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.534553 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv7gv\" (UniqueName: \"kubernetes.io/projected/3e53276a-ccb3-40a1-b1e6-307cc335ff4d-kube-api-access-dv7gv\") pod \"octavia-operator-controller-manager-6d7c7ddf95-b9jts\" (UID: \"3e53276a-ccb3-40a1-b1e6-307cc335ff4d\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-b9jts" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.534578 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktqc6\" (UniqueName: \"kubernetes.io/projected/e297e3a5-6cd9-4314-b4dc-76ba4c25c6f5-kube-api-access-ktqc6\") pod \"mariadb-operator-controller-manager-5777b4f897-j65jt\" (UID: \"e297e3a5-6cd9-4314-b4dc-76ba4c25c6f5\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-j65jt" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.534857 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-869cc7797f-7ktz6"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.534876 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-fhw55"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.535724 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-664664cb68-fhw55" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.552586 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-hbnzh" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.553317 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-mhhcx" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.562506 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-mj7x8" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.576942 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-sl5lv" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.587365 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-qnmxx"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.593230 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8zpb\" (UniqueName: \"kubernetes.io/projected/de8c01ad-9e44-42f8-956c-0bb61165a222-kube-api-access-l8zpb\") pod \"neutron-operator-controller-manager-797d478b46-cbwvq\" (UID: \"de8c01ad-9e44-42f8-956c-0bb61165a222\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-cbwvq" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.593572 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktqc6\" (UniqueName: \"kubernetes.io/projected/e297e3a5-6cd9-4314-b4dc-76ba4c25c6f5-kube-api-access-ktqc6\") pod \"mariadb-operator-controller-manager-5777b4f897-j65jt\" (UID: \"e297e3a5-6cd9-4314-b4dc-76ba4c25c6f5\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-j65jt" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.596159 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-d6gn6"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.597264 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-d6gn6" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.609052 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-n5xjf" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.626164 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-fhw55"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.635825 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-282kw\" (UniqueName: \"kubernetes.io/projected/a4e11db1-f272-469d-9c22-a649cdbcf95e-kube-api-access-282kw\") pod \"ovn-operator-controller-manager-869cc7797f-7ktz6\" (UID: \"a4e11db1-f272-469d-9c22-a649cdbcf95e\") " pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-7ktz6" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.635862 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d4c8f24c-6e84-4931-8edb-504b184ea7b0-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td\" (UID: \"d4c8f24c-6e84-4931-8edb-504b184ea7b0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.635888 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nzvg\" (UniqueName: \"kubernetes.io/projected/f1085de4-f906-4315-b8b5-a3ee4e7182c1-kube-api-access-7nzvg\") pod \"telemetry-operator-controller-manager-578874c84d-qnmxx\" (UID: \"f1085de4-f906-4315-b8b5-a3ee4e7182c1\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qnmxx" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.635928 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv7gv\" (UniqueName: \"kubernetes.io/projected/3e53276a-ccb3-40a1-b1e6-307cc335ff4d-kube-api-access-dv7gv\") pod \"octavia-operator-controller-manager-6d7c7ddf95-b9jts\" (UID: \"3e53276a-ccb3-40a1-b1e6-307cc335ff4d\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-b9jts" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.635965 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72hqs\" (UniqueName: \"kubernetes.io/projected/d363a932-2a5d-4082-841a-b5d677d0b1a1-kube-api-access-72hqs\") pod \"nova-operator-controller-manager-57bb74c7bf-ft77l\" (UID: \"d363a932-2a5d-4082-841a-b5d677d0b1a1\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-ft77l" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.641863 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddkmn\" (UniqueName: \"kubernetes.io/projected/d4c8f24c-6e84-4931-8edb-504b184ea7b0-kube-api-access-ddkmn\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td\" (UID: \"d4c8f24c-6e84-4931-8edb-504b184ea7b0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td" Oct 13 08:58:43 crc kubenswrapper[4685]: E1013 08:58:43.636113 4685 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 13 08:58:43 crc kubenswrapper[4685]: E1013 08:58:43.642397 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d4c8f24c-6e84-4931-8edb-504b184ea7b0-cert podName:d4c8f24c-6e84-4931-8edb-504b184ea7b0 nodeName:}" failed. No retries permitted until 2025-10-13 08:58:44.142367815 +0000 UTC m=+849.290243576 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d4c8f24c-6e84-4931-8edb-504b184ea7b0-cert") pod "openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td" (UID: "d4c8f24c-6e84-4931-8edb-504b184ea7b0") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.653528 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-d6gn6"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.653775 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-xf9rd" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.672429 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-cbwvq" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.692444 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-282kw\" (UniqueName: \"kubernetes.io/projected/a4e11db1-f272-469d-9c22-a649cdbcf95e-kube-api-access-282kw\") pod \"ovn-operator-controller-manager-869cc7797f-7ktz6\" (UID: \"a4e11db1-f272-469d-9c22-a649cdbcf95e\") " pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-7ktz6" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.703293 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72hqs\" (UniqueName: \"kubernetes.io/projected/d363a932-2a5d-4082-841a-b5d677d0b1a1-kube-api-access-72hqs\") pod \"nova-operator-controller-manager-57bb74c7bf-ft77l\" (UID: \"d363a932-2a5d-4082-841a-b5d677d0b1a1\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-ft77l" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.703775 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv7gv\" (UniqueName: \"kubernetes.io/projected/3e53276a-ccb3-40a1-b1e6-307cc335ff4d-kube-api-access-dv7gv\") pod \"octavia-operator-controller-manager-6d7c7ddf95-b9jts\" (UID: \"3e53276a-ccb3-40a1-b1e6-307cc335ff4d\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-b9jts" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.705485 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nzvg\" (UniqueName: \"kubernetes.io/projected/f1085de4-f906-4315-b8b5-a3ee4e7182c1-kube-api-access-7nzvg\") pod \"telemetry-operator-controller-manager-578874c84d-qnmxx\" (UID: \"f1085de4-f906-4315-b8b5-a3ee4e7182c1\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qnmxx" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.712526 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddkmn\" (UniqueName: \"kubernetes.io/projected/d4c8f24c-6e84-4931-8edb-504b184ea7b0-kube-api-access-ddkmn\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td\" (UID: \"d4c8f24c-6e84-4931-8edb-504b184ea7b0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.712582 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-dd92k"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.713554 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-dd92k" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.717234 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-rjlkf" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.731792 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-b9jts" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.735115 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-dd92k"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.749756 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mztr4\" (UniqueName: \"kubernetes.io/projected/d639bb21-69ae-45c8-8a9c-aac17f57f8dd-kube-api-access-mztr4\") pod \"swift-operator-controller-manager-5f4d5dfdc6-d6gn6\" (UID: \"d639bb21-69ae-45c8-8a9c-aac17f57f8dd\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-d6gn6" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.749818 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fflsk\" (UniqueName: \"kubernetes.io/projected/0b4f3fea-3a89-4d28-89c0-436da959e36f-kube-api-access-fflsk\") pod \"test-operator-controller-manager-ffcdd6c94-dd92k\" (UID: \"0b4f3fea-3a89-4d28-89c0-436da959e36f\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-dd92k" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.749846 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4mtf\" (UniqueName: \"kubernetes.io/projected/9239cb03-ddca-4542-b188-b89717a00f75-kube-api-access-m4mtf\") pod \"placement-operator-controller-manager-664664cb68-fhw55\" (UID: \"9239cb03-ddca-4542-b188-b89717a00f75\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-fhw55" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.749928 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a3e83b7f-5017-44e7-b507-f46a2d3f5488-cert\") pod \"infra-operator-controller-manager-585fc5b659-2bcss\" (UID: \"a3e83b7f-5017-44e7-b507-f46a2d3f5488\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-2bcss" Oct 13 08:58:43 crc kubenswrapper[4685]: E1013 08:58:43.750069 4685 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 13 08:58:43 crc kubenswrapper[4685]: E1013 08:58:43.750115 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a3e83b7f-5017-44e7-b507-f46a2d3f5488-cert podName:a3e83b7f-5017-44e7-b507-f46a2d3f5488 nodeName:}" failed. No retries permitted until 2025-10-13 08:58:44.750100566 +0000 UTC m=+849.897976327 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a3e83b7f-5017-44e7-b507-f46a2d3f5488-cert") pod "infra-operator-controller-manager-585fc5b659-2bcss" (UID: "a3e83b7f-5017-44e7-b507-f46a2d3f5488") : secret "infra-operator-webhook-server-cert" not found Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.753213 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-9wdvd"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.754488 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-646675d848-9wdvd" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.776972 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-q6smr" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.805488 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-9wdvd"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.806167 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-7ktz6" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.835552 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qnmxx" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.854715 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fflsk\" (UniqueName: \"kubernetes.io/projected/0b4f3fea-3a89-4d28-89c0-436da959e36f-kube-api-access-fflsk\") pod \"test-operator-controller-manager-ffcdd6c94-dd92k\" (UID: \"0b4f3fea-3a89-4d28-89c0-436da959e36f\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-dd92k" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.860092 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4mtf\" (UniqueName: \"kubernetes.io/projected/9239cb03-ddca-4542-b188-b89717a00f75-kube-api-access-m4mtf\") pod \"placement-operator-controller-manager-664664cb68-fhw55\" (UID: \"9239cb03-ddca-4542-b188-b89717a00f75\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-fhw55" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.860226 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ggpc\" (UniqueName: \"kubernetes.io/projected/026b2615-8dc1-4ba7-83d1-1e21f4fa80d2-kube-api-access-7ggpc\") pod \"watcher-operator-controller-manager-646675d848-9wdvd\" (UID: \"026b2615-8dc1-4ba7-83d1-1e21f4fa80d2\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-9wdvd" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.860394 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mztr4\" (UniqueName: \"kubernetes.io/projected/d639bb21-69ae-45c8-8a9c-aac17f57f8dd-kube-api-access-mztr4\") pod \"swift-operator-controller-manager-5f4d5dfdc6-d6gn6\" (UID: \"d639bb21-69ae-45c8-8a9c-aac17f57f8dd\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-d6gn6" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.880168 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-j65jt" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.903435 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4mtf\" (UniqueName: \"kubernetes.io/projected/9239cb03-ddca-4542-b188-b89717a00f75-kube-api-access-m4mtf\") pod \"placement-operator-controller-manager-664664cb68-fhw55\" (UID: \"9239cb03-ddca-4542-b188-b89717a00f75\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-fhw55" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.906740 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-859d8f6f74-jwkhb"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.911360 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-859d8f6f74-jwkhb" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.913526 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fflsk\" (UniqueName: \"kubernetes.io/projected/0b4f3fea-3a89-4d28-89c0-436da959e36f-kube-api-access-fflsk\") pod \"test-operator-controller-manager-ffcdd6c94-dd92k\" (UID: \"0b4f3fea-3a89-4d28-89c0-436da959e36f\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-dd92k" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.926103 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-859d8f6f74-jwkhb"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.926242 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.926349 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-sqwv6" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.928524 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bgvk8"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.929454 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bgvk8" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.932805 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mztr4\" (UniqueName: \"kubernetes.io/projected/d639bb21-69ae-45c8-8a9c-aac17f57f8dd-kube-api-access-mztr4\") pod \"swift-operator-controller-manager-5f4d5dfdc6-d6gn6\" (UID: \"d639bb21-69ae-45c8-8a9c-aac17f57f8dd\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-d6gn6" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.933114 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-8n9br" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.934163 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bgvk8"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.934571 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-664664cb68-fhw55" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.943454 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-zm9bd"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.951233 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-5d7zf"] Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.955263 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-d6gn6" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.961706 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ggpc\" (UniqueName: \"kubernetes.io/projected/026b2615-8dc1-4ba7-83d1-1e21f4fa80d2-kube-api-access-7ggpc\") pod \"watcher-operator-controller-manager-646675d848-9wdvd\" (UID: \"026b2615-8dc1-4ba7-83d1-1e21f4fa80d2\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-9wdvd" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.961792 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jtml\" (UniqueName: \"kubernetes.io/projected/d2886500-f092-4cba-bd20-50483b96ceb3-kube-api-access-6jtml\") pod \"openstack-operator-controller-manager-859d8f6f74-jwkhb\" (UID: \"d2886500-f092-4cba-bd20-50483b96ceb3\") " pod="openstack-operators/openstack-operator-controller-manager-859d8f6f74-jwkhb" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.961824 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d2886500-f092-4cba-bd20-50483b96ceb3-cert\") pod \"openstack-operator-controller-manager-859d8f6f74-jwkhb\" (UID: \"d2886500-f092-4cba-bd20-50483b96ceb3\") " pod="openstack-operators/openstack-operator-controller-manager-859d8f6f74-jwkhb" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.961884 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzrcs\" (UniqueName: \"kubernetes.io/projected/79c6993d-5a5b-4b38-b678-6c78e639d834-kube-api-access-bzrcs\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-bgvk8\" (UID: \"79c6993d-5a5b-4b38-b678-6c78e639d834\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bgvk8" Oct 13 08:58:43 crc kubenswrapper[4685]: I1013 08:58:43.974927 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-ft77l" Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.008966 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ggpc\" (UniqueName: \"kubernetes.io/projected/026b2615-8dc1-4ba7-83d1-1e21f4fa80d2-kube-api-access-7ggpc\") pod \"watcher-operator-controller-manager-646675d848-9wdvd\" (UID: \"026b2615-8dc1-4ba7-83d1-1e21f4fa80d2\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-9wdvd" Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.078945 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jtml\" (UniqueName: \"kubernetes.io/projected/d2886500-f092-4cba-bd20-50483b96ceb3-kube-api-access-6jtml\") pod \"openstack-operator-controller-manager-859d8f6f74-jwkhb\" (UID: \"d2886500-f092-4cba-bd20-50483b96ceb3\") " pod="openstack-operators/openstack-operator-controller-manager-859d8f6f74-jwkhb" Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.078994 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d2886500-f092-4cba-bd20-50483b96ceb3-cert\") pod \"openstack-operator-controller-manager-859d8f6f74-jwkhb\" (UID: \"d2886500-f092-4cba-bd20-50483b96ceb3\") " pod="openstack-operators/openstack-operator-controller-manager-859d8f6f74-jwkhb" Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.079053 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzrcs\" (UniqueName: \"kubernetes.io/projected/79c6993d-5a5b-4b38-b678-6c78e639d834-kube-api-access-bzrcs\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-bgvk8\" (UID: \"79c6993d-5a5b-4b38-b678-6c78e639d834\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bgvk8" Oct 13 08:58:44 crc kubenswrapper[4685]: E1013 08:58:44.079508 4685 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 13 08:58:44 crc kubenswrapper[4685]: E1013 08:58:44.079544 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d2886500-f092-4cba-bd20-50483b96ceb3-cert podName:d2886500-f092-4cba-bd20-50483b96ceb3 nodeName:}" failed. No retries permitted until 2025-10-13 08:58:44.5795312 +0000 UTC m=+849.727406961 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d2886500-f092-4cba-bd20-50483b96ceb3-cert") pod "openstack-operator-controller-manager-859d8f6f74-jwkhb" (UID: "d2886500-f092-4cba-bd20-50483b96ceb3") : secret "webhook-server-cert" not found Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.089318 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-dd92k" Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.113144 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jtml\" (UniqueName: \"kubernetes.io/projected/d2886500-f092-4cba-bd20-50483b96ceb3-kube-api-access-6jtml\") pod \"openstack-operator-controller-manager-859d8f6f74-jwkhb\" (UID: \"d2886500-f092-4cba-bd20-50483b96ceb3\") " pod="openstack-operators/openstack-operator-controller-manager-859d8f6f74-jwkhb" Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.123602 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzrcs\" (UniqueName: \"kubernetes.io/projected/79c6993d-5a5b-4b38-b678-6c78e639d834-kube-api-access-bzrcs\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-bgvk8\" (UID: \"79c6993d-5a5b-4b38-b678-6c78e639d834\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bgvk8" Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.141536 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-646675d848-9wdvd" Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.187596 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d4c8f24c-6e84-4931-8edb-504b184ea7b0-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td\" (UID: \"d4c8f24c-6e84-4931-8edb-504b184ea7b0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td" Oct 13 08:58:44 crc kubenswrapper[4685]: E1013 08:58:44.187836 4685 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 13 08:58:44 crc kubenswrapper[4685]: E1013 08:58:44.187883 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d4c8f24c-6e84-4931-8edb-504b184ea7b0-cert podName:d4c8f24c-6e84-4931-8edb-504b184ea7b0 nodeName:}" failed. No retries permitted until 2025-10-13 08:58:45.187870389 +0000 UTC m=+850.335746150 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d4c8f24c-6e84-4931-8edb-504b184ea7b0-cert") pod "openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td" (UID: "d4c8f24c-6e84-4931-8edb-504b184ea7b0") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.312642 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-qbl58"] Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.321276 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bgvk8" Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.432560 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp"] Oct 13 08:58:44 crc kubenswrapper[4685]: W1013 08:58:44.454909 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc92c1fba_b02b_4b6c_9570_b75ee60c5e86.slice/crio-f9e227e1605c8c5db8dcc87d0e2c647f4420d5f9f8b2a7604d6d77d9a0bb88a1 WatchSource:0}: Error finding container f9e227e1605c8c5db8dcc87d0e2c647f4420d5f9f8b2a7604d6d77d9a0bb88a1: Status 404 returned error can't find the container with id f9e227e1605c8c5db8dcc87d0e2c647f4420d5f9f8b2a7604d6d77d9a0bb88a1 Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.486819 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-hbnzh"] Oct 13 08:58:44 crc kubenswrapper[4685]: W1013 08:58:44.531007 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb001d17a_1aea_44ba_86c5_ba6b312156c1.slice/crio-f5f8e1231912691c1bcae211244492c05819f3f9c9b295126ad628a27439b874 WatchSource:0}: Error finding container f5f8e1231912691c1bcae211244492c05819f3f9c9b295126ad628a27439b874: Status 404 returned error can't find the container with id f5f8e1231912691c1bcae211244492c05819f3f9c9b295126ad628a27439b874 Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.600970 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d2886500-f092-4cba-bd20-50483b96ceb3-cert\") pod \"openstack-operator-controller-manager-859d8f6f74-jwkhb\" (UID: \"d2886500-f092-4cba-bd20-50483b96ceb3\") " pod="openstack-operators/openstack-operator-controller-manager-859d8f6f74-jwkhb" Oct 13 08:58:44 crc kubenswrapper[4685]: E1013 08:58:44.601272 4685 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 13 08:58:44 crc kubenswrapper[4685]: E1013 08:58:44.601325 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d2886500-f092-4cba-bd20-50483b96ceb3-cert podName:d2886500-f092-4cba-bd20-50483b96ceb3 nodeName:}" failed. No retries permitted until 2025-10-13 08:58:45.601310504 +0000 UTC m=+850.749186265 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d2886500-f092-4cba-bd20-50483b96ceb3-cert") pod "openstack-operator-controller-manager-859d8f6f74-jwkhb" (UID: "d2886500-f092-4cba-bd20-50483b96ceb3") : secret "webhook-server-cert" not found Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.684508 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-q8dkt"] Oct 13 08:58:44 crc kubenswrapper[4685]: W1013 08:58:44.717214 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12dc02ce_76d4_4376_ab6f_b2a75580cc4d.slice/crio-77e6a86b5e7b60c6fa87e27b2087f0532d555ca1ba226dedaae808b264c25e32 WatchSource:0}: Error finding container 77e6a86b5e7b60c6fa87e27b2087f0532d555ca1ba226dedaae808b264c25e32: Status 404 returned error can't find the container with id 77e6a86b5e7b60c6fa87e27b2087f0532d555ca1ba226dedaae808b264c25e32 Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.745235 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-xf9rd"] Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.746246 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-5d7zf" event={"ID":"e509e801-67de-4a55-bd22-cf3f73deca81","Type":"ContainerStarted","Data":"3bb7bfc02d9c4e5ffd9ae38ef22db3be42c91c6916379c28d929bdf6b10422ff"} Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.760609 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-sl5lv"] Oct 13 08:58:44 crc kubenswrapper[4685]: W1013 08:58:44.770573 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb08316b_7fd3_4d65_88e1_bbc91efcb7c7.slice/crio-8f49932a0563b5c9558acad6d291f88381d00a6ee82959fca04199f4f478c533 WatchSource:0}: Error finding container 8f49932a0563b5c9558acad6d291f88381d00a6ee82959fca04199f4f478c533: Status 404 returned error can't find the container with id 8f49932a0563b5c9558acad6d291f88381d00a6ee82959fca04199f4f478c533 Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.770726 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" event={"ID":"b001d17a-1aea-44ba-86c5-ba6b312156c1","Type":"ContainerStarted","Data":"f5f8e1231912691c1bcae211244492c05819f3f9c9b295126ad628a27439b874"} Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.778819 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zm9bd" event={"ID":"54405c7c-61f7-41ea-ae0a-29128b51326c","Type":"ContainerStarted","Data":"572088f07448cc5c9122f3ec0c7f23851d5f2c20286cfa2cb60f7dd8cb9d8187"} Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.785873 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-qbl58" event={"ID":"c92c1fba-b02b-4b6c-9570-b75ee60c5e86","Type":"ContainerStarted","Data":"f9e227e1605c8c5db8dcc87d0e2c647f4420d5f9f8b2a7604d6d77d9a0bb88a1"} Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.787831 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-hbnzh" event={"ID":"92a042d7-669e-48d2-8d7e-8a8da4fc01eb","Type":"ContainerStarted","Data":"cfd782cd1d8af1b444f32126d603a1098c0a3a395daf33046865d71cc9eb58d9"} Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.805444 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a3e83b7f-5017-44e7-b507-f46a2d3f5488-cert\") pod \"infra-operator-controller-manager-585fc5b659-2bcss\" (UID: \"a3e83b7f-5017-44e7-b507-f46a2d3f5488\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-2bcss" Oct 13 08:58:44 crc kubenswrapper[4685]: I1013 08:58:44.809658 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a3e83b7f-5017-44e7-b507-f46a2d3f5488-cert\") pod \"infra-operator-controller-manager-585fc5b659-2bcss\" (UID: \"a3e83b7f-5017-44e7-b507-f46a2d3f5488\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-2bcss" Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.033824 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-qnmxx"] Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.046246 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-b9jts"] Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.051418 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-cbwvq"] Oct 13 08:58:45 crc kubenswrapper[4685]: W1013 08:58:45.056137 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e53276a_ccb3_40a1_b1e6_307cc335ff4d.slice/crio-7c7132167fa8a1ac739a41fc4ce9ad9cccf445838012a20f481852302e75064e WatchSource:0}: Error finding container 7c7132167fa8a1ac739a41fc4ce9ad9cccf445838012a20f481852302e75064e: Status 404 returned error can't find the container with id 7c7132167fa8a1ac739a41fc4ce9ad9cccf445838012a20f481852302e75064e Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.058816 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-mj7x8"] Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.102880 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-2bcss" Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.210012 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d4c8f24c-6e84-4931-8edb-504b184ea7b0-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td\" (UID: \"d4c8f24c-6e84-4931-8edb-504b184ea7b0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td" Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.228576 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d4c8f24c-6e84-4931-8edb-504b184ea7b0-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td\" (UID: \"d4c8f24c-6e84-4931-8edb-504b184ea7b0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td" Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.280289 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td" Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.430904 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-j65jt"] Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.446128 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-dd92k"] Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.456145 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-d6gn6"] Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.466354 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-869cc7797f-7ktz6"] Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.472740 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-ft77l"] Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.488982 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bgvk8"] Oct 13 08:58:45 crc kubenswrapper[4685]: E1013 08:58:45.512632 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m4mtf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-664664cb68-fhw55_openstack-operators(9239cb03-ddca-4542-b188-b89717a00f75): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 13 08:58:45 crc kubenswrapper[4685]: E1013 08:58:45.512847 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mztr4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f4d5dfdc6-d6gn6_openstack-operators(d639bb21-69ae-45c8-8a9c-aac17f57f8dd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 13 08:58:45 crc kubenswrapper[4685]: E1013 08:58:45.512993 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7ggpc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-646675d848-9wdvd_openstack-operators(026b2615-8dc1-4ba7-83d1-1e21f4fa80d2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.587664 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-fhw55"] Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.593700 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-9wdvd"] Oct 13 08:58:45 crc kubenswrapper[4685]: W1013 08:58:45.606286 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3e83b7f_5017_44e7_b507_f46a2d3f5488.slice/crio-6d9a40941960c67450c7eb8194bee3e11138aece05c408e25f949be0d0cdaa1d WatchSource:0}: Error finding container 6d9a40941960c67450c7eb8194bee3e11138aece05c408e25f949be0d0cdaa1d: Status 404 returned error can't find the container with id 6d9a40941960c67450c7eb8194bee3e11138aece05c408e25f949be0d0cdaa1d Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.612278 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-2bcss"] Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.615887 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d2886500-f092-4cba-bd20-50483b96ceb3-cert\") pod \"openstack-operator-controller-manager-859d8f6f74-jwkhb\" (UID: \"d2886500-f092-4cba-bd20-50483b96ceb3\") " pod="openstack-operators/openstack-operator-controller-manager-859d8f6f74-jwkhb" Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.626221 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d2886500-f092-4cba-bd20-50483b96ceb3-cert\") pod \"openstack-operator-controller-manager-859d8f6f74-jwkhb\" (UID: \"d2886500-f092-4cba-bd20-50483b96ceb3\") " pod="openstack-operators/openstack-operator-controller-manager-859d8f6f74-jwkhb" Oct 13 08:58:45 crc kubenswrapper[4685]: E1013 08:58:45.643222 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d7nx6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-585fc5b659-2bcss_openstack-operators(a3e83b7f-5017-44e7-b507-f46a2d3f5488): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 13 08:58:45 crc kubenswrapper[4685]: E1013 08:58:45.747694 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-d6gn6" podUID="d639bb21-69ae-45c8-8a9c-aac17f57f8dd" Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.785748 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-859d8f6f74-jwkhb" Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.803906 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td"] Oct 13 08:58:45 crc kubenswrapper[4685]: W1013 08:58:45.825902 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4c8f24c_6e84_4931_8edb_504b184ea7b0.slice/crio-9d9197eb7e5ceec17200c8392140f754bb4d106eb99e37d24add977c408a8be9 WatchSource:0}: Error finding container 9d9197eb7e5ceec17200c8392140f754bb4d106eb99e37d24add977c408a8be9: Status 404 returned error can't find the container with id 9d9197eb7e5ceec17200c8392140f754bb4d106eb99e37d24add977c408a8be9 Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.842463 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-b9jts" event={"ID":"3e53276a-ccb3-40a1-b1e6-307cc335ff4d","Type":"ContainerStarted","Data":"7c7132167fa8a1ac739a41fc4ce9ad9cccf445838012a20f481852302e75064e"} Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.851439 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-2bcss" event={"ID":"a3e83b7f-5017-44e7-b507-f46a2d3f5488","Type":"ContainerStarted","Data":"6d9a40941960c67450c7eb8194bee3e11138aece05c408e25f949be0d0cdaa1d"} Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.852693 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bgvk8" event={"ID":"79c6993d-5a5b-4b38-b678-6c78e639d834","Type":"ContainerStarted","Data":"d0a9e5ba0ceb59b22ff5cf1c8e4e7a34d5563805c358a8aae57232c8e44ba81c"} Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.862926 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-d6gn6" event={"ID":"d639bb21-69ae-45c8-8a9c-aac17f57f8dd","Type":"ContainerStarted","Data":"a04c73f75d52d69075ddca536c9c28f93be0fd597d605a05de3bb3724c260b66"} Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.863012 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-d6gn6" event={"ID":"d639bb21-69ae-45c8-8a9c-aac17f57f8dd","Type":"ContainerStarted","Data":"52a4cb4dc55cc2ed5c54901645e3d1a0d48ae9ec051451aa345af95f844b37d2"} Oct 13 08:58:45 crc kubenswrapper[4685]: E1013 08:58:45.865336 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-d6gn6" podUID="d639bb21-69ae-45c8-8a9c-aac17f57f8dd" Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.872754 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-fhw55" event={"ID":"9239cb03-ddca-4542-b188-b89717a00f75","Type":"ContainerStarted","Data":"ec7f236440ab1bb0f26bad4d96fed98384dbd1e5aa75863df589a6387fcffe6c"} Oct 13 08:58:45 crc kubenswrapper[4685]: E1013 08:58:45.875455 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-fhw55" podUID="9239cb03-ddca-4542-b188-b89717a00f75" Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.876120 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-mj7x8" event={"ID":"6b61f732-3f39-459e-bec0-a6bdd5added4","Type":"ContainerStarted","Data":"5494521d7e3c158b61d8111684dd184a032ec9466aa444ebd8f608c18824782a"} Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.904084 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-xf9rd" event={"ID":"9899b0d7-81a4-49f7-91cd-3c5aa72d49b2","Type":"ContainerStarted","Data":"bb64ee06567b7fc8423659a600fbcbb987f020730dbb882b05ee2b4adde07b90"} Oct 13 08:58:45 crc kubenswrapper[4685]: E1013 08:58:45.919024 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-9wdvd" podUID="026b2615-8dc1-4ba7-83d1-1e21f4fa80d2" Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.919491 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-dd92k" event={"ID":"0b4f3fea-3a89-4d28-89c0-436da959e36f","Type":"ContainerStarted","Data":"061fa2d5761329c449d9b641f043265eec74a8b732d90bd873c0180355fae80d"} Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.920641 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-sl5lv" event={"ID":"eb08316b-7fd3-4d65-88e1-bbc91efcb7c7","Type":"ContainerStarted","Data":"8f49932a0563b5c9558acad6d291f88381d00a6ee82959fca04199f4f478c533"} Oct 13 08:58:45 crc kubenswrapper[4685]: E1013 08:58:45.934117 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-2bcss" podUID="a3e83b7f-5017-44e7-b507-f46a2d3f5488" Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.939985 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-7ktz6" event={"ID":"a4e11db1-f272-469d-9c22-a649cdbcf95e","Type":"ContainerStarted","Data":"b608254612bbed1d43781cb34d093c5a909a1abf97155c0205ccbbe46aa504b2"} Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.942875 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qnmxx" event={"ID":"f1085de4-f906-4315-b8b5-a3ee4e7182c1","Type":"ContainerStarted","Data":"7bcf55f4af43be9634b33af76d8b289b12dad84234309d74ad013ab1dcc6711d"} Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.944381 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-9wdvd" event={"ID":"026b2615-8dc1-4ba7-83d1-1e21f4fa80d2","Type":"ContainerStarted","Data":"866e99ebcc3249d110b659283d91370309e0aaa67a1f4b7da89a0b1479e6c3c9"} Oct 13 08:58:45 crc kubenswrapper[4685]: E1013 08:58:45.946006 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-9wdvd" podUID="026b2615-8dc1-4ba7-83d1-1e21f4fa80d2" Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.951937 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-j65jt" event={"ID":"e297e3a5-6cd9-4314-b4dc-76ba4c25c6f5","Type":"ContainerStarted","Data":"e385ccad40bbe73c99acb87b0d3389174d0b440536fe6b902743c7f4f08b3b67"} Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.958515 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-cbwvq" event={"ID":"de8c01ad-9e44-42f8-956c-0bb61165a222","Type":"ContainerStarted","Data":"1c5fb58ea3671205b06ba8138d9ffe91402c43dfa8f870a26f8aa18add3a6315"} Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.974680 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-q8dkt" event={"ID":"12dc02ce-76d4-4376-ab6f-b2a75580cc4d","Type":"ContainerStarted","Data":"77e6a86b5e7b60c6fa87e27b2087f0532d555ca1ba226dedaae808b264c25e32"} Oct 13 08:58:45 crc kubenswrapper[4685]: I1013 08:58:45.994477 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-ft77l" event={"ID":"d363a932-2a5d-4082-841a-b5d677d0b1a1","Type":"ContainerStarted","Data":"bc9f2a1b1208c4d847b2dc8a0b338298371bfe679b15da596a10a9401d46a097"} Oct 13 08:58:46 crc kubenswrapper[4685]: I1013 08:58:46.467740 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-859d8f6f74-jwkhb"] Oct 13 08:58:47 crc kubenswrapper[4685]: I1013 08:58:47.013533 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-2bcss" event={"ID":"a3e83b7f-5017-44e7-b507-f46a2d3f5488","Type":"ContainerStarted","Data":"7ad980b1875b386780c018069acb102c40a8daf36b852a81706abd0448a6e4bf"} Oct 13 08:58:47 crc kubenswrapper[4685]: E1013 08:58:47.019535 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492\\\"\"" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-2bcss" podUID="a3e83b7f-5017-44e7-b507-f46a2d3f5488" Oct 13 08:58:47 crc kubenswrapper[4685]: I1013 08:58:47.025824 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-859d8f6f74-jwkhb" event={"ID":"d2886500-f092-4cba-bd20-50483b96ceb3","Type":"ContainerStarted","Data":"4941eb95fd82acee90545b4d7d07701fd1f1654739902d22d1e19fae2eb9f4b2"} Oct 13 08:58:47 crc kubenswrapper[4685]: I1013 08:58:47.025865 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-859d8f6f74-jwkhb" event={"ID":"d2886500-f092-4cba-bd20-50483b96ceb3","Type":"ContainerStarted","Data":"241297200dc2be2b11825f7df988d4a0e81f196debf747e30a7df112bf85ffc6"} Oct 13 08:58:47 crc kubenswrapper[4685]: I1013 08:58:47.025878 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-859d8f6f74-jwkhb" event={"ID":"d2886500-f092-4cba-bd20-50483b96ceb3","Type":"ContainerStarted","Data":"b7295fdcd511b24bb2846eb50cc87380e791d0fd9da259565509f9afce4b1966"} Oct 13 08:58:47 crc kubenswrapper[4685]: I1013 08:58:47.026656 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-859d8f6f74-jwkhb" Oct 13 08:58:47 crc kubenswrapper[4685]: I1013 08:58:47.039892 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-9wdvd" event={"ID":"026b2615-8dc1-4ba7-83d1-1e21f4fa80d2","Type":"ContainerStarted","Data":"c80eda5255dc2e95110e9a15d7468ba5191dfaa8900a72910661115a74484249"} Oct 13 08:58:47 crc kubenswrapper[4685]: E1013 08:58:47.042105 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-9wdvd" podUID="026b2615-8dc1-4ba7-83d1-1e21f4fa80d2" Oct 13 08:58:47 crc kubenswrapper[4685]: I1013 08:58:47.050315 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td" event={"ID":"d4c8f24c-6e84-4931-8edb-504b184ea7b0","Type":"ContainerStarted","Data":"9d9197eb7e5ceec17200c8392140f754bb4d106eb99e37d24add977c408a8be9"} Oct 13 08:58:47 crc kubenswrapper[4685]: I1013 08:58:47.067791 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-fhw55" event={"ID":"9239cb03-ddca-4542-b188-b89717a00f75","Type":"ContainerStarted","Data":"dcb445dcdfb6c6ddbd0fd41a54b472dca764055725609c0c8e8312eba5bb4876"} Oct 13 08:58:47 crc kubenswrapper[4685]: E1013 08:58:47.070293 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-d6gn6" podUID="d639bb21-69ae-45c8-8a9c-aac17f57f8dd" Oct 13 08:58:47 crc kubenswrapper[4685]: E1013 08:58:47.070732 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-fhw55" podUID="9239cb03-ddca-4542-b188-b89717a00f75" Oct 13 08:58:47 crc kubenswrapper[4685]: I1013 08:58:47.104697 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-859d8f6f74-jwkhb" podStartSLOduration=4.104677572 podStartE2EDuration="4.104677572s" podCreationTimestamp="2025-10-13 08:58:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 08:58:47.087603235 +0000 UTC m=+852.235478996" watchObservedRunningTime="2025-10-13 08:58:47.104677572 +0000 UTC m=+852.252553333" Oct 13 08:58:48 crc kubenswrapper[4685]: E1013 08:58:48.081099 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-9wdvd" podUID="026b2615-8dc1-4ba7-83d1-1e21f4fa80d2" Oct 13 08:58:48 crc kubenswrapper[4685]: E1013 08:58:48.081189 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-fhw55" podUID="9239cb03-ddca-4542-b188-b89717a00f75" Oct 13 08:58:48 crc kubenswrapper[4685]: E1013 08:58:48.082746 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492\\\"\"" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-2bcss" podUID="a3e83b7f-5017-44e7-b507-f46a2d3f5488" Oct 13 08:58:52 crc kubenswrapper[4685]: I1013 08:58:52.980707 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 08:58:52 crc kubenswrapper[4685]: I1013 08:58:52.981198 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 08:58:52 crc kubenswrapper[4685]: I1013 08:58:52.981258 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 08:58:52 crc kubenswrapper[4685]: I1013 08:58:52.982202 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c54513235e556be91a1895a9b0943234dc7b6e87cbf946c33f9dc14ea2f61818"} pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 08:58:52 crc kubenswrapper[4685]: I1013 08:58:52.982300 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" containerID="cri-o://c54513235e556be91a1895a9b0943234dc7b6e87cbf946c33f9dc14ea2f61818" gracePeriod=600 Oct 13 08:58:54 crc kubenswrapper[4685]: I1013 08:58:54.120083 4685 generic.go:334] "Generic (PLEG): container finished" podID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerID="c54513235e556be91a1895a9b0943234dc7b6e87cbf946c33f9dc14ea2f61818" exitCode=0 Oct 13 08:58:54 crc kubenswrapper[4685]: I1013 08:58:54.120147 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerDied","Data":"c54513235e556be91a1895a9b0943234dc7b6e87cbf946c33f9dc14ea2f61818"} Oct 13 08:58:54 crc kubenswrapper[4685]: I1013 08:58:54.120444 4685 scope.go:117] "RemoveContainer" containerID="c37641321127c1e407fb4df4bcfeb42705776322c36634a9a7d0dc9c4a308038" Oct 13 08:58:55 crc kubenswrapper[4685]: I1013 08:58:55.791382 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-859d8f6f74-jwkhb" Oct 13 08:59:01 crc kubenswrapper[4685]: E1013 08:59:01.532733 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2" Oct 13 08:59:01 crc kubenswrapper[4685]: E1013 08:59:01.533666 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l8zpb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-797d478b46-cbwvq_openstack-operators(de8c01ad-9e44-42f8-956c-0bb61165a222): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 08:59:01 crc kubenswrapper[4685]: E1013 08:59:01.649817 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.174:5001/openstack-k8s-operators/keystone-operator:5d735e29a58b1e4afb30ec827263ac7cab51ba01" Oct 13 08:59:01 crc kubenswrapper[4685]: E1013 08:59:01.649969 4685 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.174:5001/openstack-k8s-operators/keystone-operator:5d735e29a58b1e4afb30ec827263ac7cab51ba01" Oct 13 08:59:01 crc kubenswrapper[4685]: E1013 08:59:01.650441 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.174:5001/openstack-k8s-operators/keystone-operator:5d735e29a58b1e4afb30ec827263ac7cab51ba01,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cqbtq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 08:59:02 crc kubenswrapper[4685]: E1013 08:59:02.193858 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:09deecf840d38ff6af3c924729cf0a9444bc985848bfbe7c918019b88a6bc4d7" Oct 13 08:59:02 crc kubenswrapper[4685]: E1013 08:59:02.194384 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:09deecf840d38ff6af3c924729cf0a9444bc985848bfbe7c918019b88a6bc4d7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dv7gv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-6d7c7ddf95-b9jts_openstack-operators(3e53276a-ccb3-40a1-b1e6-307cc335ff4d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 08:59:02 crc kubenswrapper[4685]: E1013 08:59:02.629031 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:7e584b1c430441c8b6591dadeff32e065de8a185ad37ef90d2e08d37e59aab4a" Oct 13 08:59:02 crc kubenswrapper[4685]: E1013 08:59:02.629186 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:7e584b1c430441c8b6591dadeff32e065de8a185ad37ef90d2e08d37e59aab4a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fflsk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-ffcdd6c94-dd92k_openstack-operators(0b4f3fea-3a89-4d28-89c0-436da959e36f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 08:59:03 crc kubenswrapper[4685]: E1013 08:59:03.745456 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:783f711b4cb179819cfcb81167c3591c70671440f4551bbe48b7a8730567f577" Oct 13 08:59:03 crc kubenswrapper[4685]: E1013 08:59:03.745626 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:783f711b4cb179819cfcb81167c3591c70671440f4551bbe48b7a8730567f577,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wwkjl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-64f84fcdbb-5d7zf_openstack-operators(e509e801-67de-4a55-bd22-cf3f73deca81): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 08:59:04 crc kubenswrapper[4685]: E1013 08:59:04.829973 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:063a7e65b4ba98f0506f269ff7525b446eae06a5ed4a61c18ffa33a886500867" Oct 13 08:59:04 crc kubenswrapper[4685]: E1013 08:59:04.830496 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:063a7e65b4ba98f0506f269ff7525b446eae06a5ed4a61c18ffa33a886500867,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q7dbl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-6d74794d9b-sl5lv_openstack-operators(eb08316b-7fd3-4d65-88e1-bbc91efcb7c7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 08:59:05 crc kubenswrapper[4685]: E1013 08:59:05.918197 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14" Oct 13 08:59:05 crc kubenswrapper[4685]: E1013 08:59:05.918411 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-282kw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-869cc7797f-7ktz6_openstack-operators(a4e11db1-f272-469d-9c22-a649cdbcf95e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 08:59:06 crc kubenswrapper[4685]: E1013 08:59:06.444636 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:3cc6bba71197ddf88dd4ba1301542bacbc1fe12e6faab2b69e6960944b3d74a0" Oct 13 08:59:06 crc kubenswrapper[4685]: E1013 08:59:06.444848 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:3cc6bba71197ddf88dd4ba1301542bacbc1fe12e6faab2b69e6960944b3d74a0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xfj4l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-7bb46cd7d-zm9bd_openstack-operators(54405c7c-61f7-41ea-ae0a-29128b51326c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 08:59:09 crc kubenswrapper[4685]: E1013 08:59:09.081361 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-cbwvq" podUID="de8c01ad-9e44-42f8-956c-0bb61165a222" Oct 13 08:59:09 crc kubenswrapper[4685]: E1013 08:59:09.130970 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-dd92k" podUID="0b4f3fea-3a89-4d28-89c0-436da959e36f" Oct 13 08:59:09 crc kubenswrapper[4685]: E1013 08:59:09.138737 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-5d7zf" podUID="e509e801-67de-4a55-bd22-cf3f73deca81" Oct 13 08:59:09 crc kubenswrapper[4685]: E1013 08:59:09.169013 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-b9jts" podUID="3e53276a-ccb3-40a1-b1e6-307cc335ff4d" Oct 13 08:59:09 crc kubenswrapper[4685]: E1013 08:59:09.238273 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-sl5lv" podUID="eb08316b-7fd3-4d65-88e1-bbc91efcb7c7" Oct 13 08:59:09 crc kubenswrapper[4685]: E1013 08:59:09.239556 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 08:59:09 crc kubenswrapper[4685]: I1013 08:59:09.272145 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" event={"ID":"b001d17a-1aea-44ba-86c5-ba6b312156c1","Type":"ContainerStarted","Data":"d649ef2bd28c9a0e294ae79299a01922ef6d50178891e5086d99bd62659679ed"} Oct 13 08:59:09 crc kubenswrapper[4685]: I1013 08:59:09.349529 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-sl5lv" event={"ID":"eb08316b-7fd3-4d65-88e1-bbc91efcb7c7","Type":"ContainerStarted","Data":"2aa9d987229c5652c533b6e9a741e750b87f31b41ae8912b38394f908699afda"} Oct 13 08:59:09 crc kubenswrapper[4685]: E1013 08:59:09.359561 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.174:5001/openstack-k8s-operators/keystone-operator:5d735e29a58b1e4afb30ec827263ac7cab51ba01\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 08:59:09 crc kubenswrapper[4685]: I1013 08:59:09.359829 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerStarted","Data":"722c93d4dbc8e9002b13fe5143e24de1b4add33c78bcd87a264b498be58dd6fd"} Oct 13 08:59:09 crc kubenswrapper[4685]: E1013 08:59:09.360158 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:063a7e65b4ba98f0506f269ff7525b446eae06a5ed4a61c18ffa33a886500867\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-sl5lv" podUID="eb08316b-7fd3-4d65-88e1-bbc91efcb7c7" Oct 13 08:59:09 crc kubenswrapper[4685]: I1013 08:59:09.375941 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-dd92k" event={"ID":"0b4f3fea-3a89-4d28-89c0-436da959e36f","Type":"ContainerStarted","Data":"68aa219b06aa4624d486d42fc94ee4d4fa679b71f5fbedebd0b000171a7e36b1"} Oct 13 08:59:09 crc kubenswrapper[4685]: E1013 08:59:09.380969 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:7e584b1c430441c8b6591dadeff32e065de8a185ad37ef90d2e08d37e59aab4a\\\"\"" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-dd92k" podUID="0b4f3fea-3a89-4d28-89c0-436da959e36f" Oct 13 08:59:09 crc kubenswrapper[4685]: I1013 08:59:09.410181 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-5d7zf" event={"ID":"e509e801-67de-4a55-bd22-cf3f73deca81","Type":"ContainerStarted","Data":"8f5c5491392502a77f966259aad2c8aa665f844e1d0599fea6affcd98f951013"} Oct 13 08:59:09 crc kubenswrapper[4685]: E1013 08:59:09.417128 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:783f711b4cb179819cfcb81167c3591c70671440f4551bbe48b7a8730567f577\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-5d7zf" podUID="e509e801-67de-4a55-bd22-cf3f73deca81" Oct 13 08:59:09 crc kubenswrapper[4685]: I1013 08:59:09.426074 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bgvk8" event={"ID":"79c6993d-5a5b-4b38-b678-6c78e639d834","Type":"ContainerStarted","Data":"2b37b86573fa52d89e87b71a7f523b23a8f414d41a008a09b03e35071d862d7c"} Oct 13 08:59:09 crc kubenswrapper[4685]: I1013 08:59:09.444117 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-cbwvq" event={"ID":"de8c01ad-9e44-42f8-956c-0bb61165a222","Type":"ContainerStarted","Data":"1ba931b254aaf0a66ab8567830008b18e7acf790e9647a605bbcc3d33abf4cbc"} Oct 13 08:59:09 crc kubenswrapper[4685]: E1013 08:59:09.450223 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-cbwvq" podUID="de8c01ad-9e44-42f8-956c-0bb61165a222" Oct 13 08:59:09 crc kubenswrapper[4685]: I1013 08:59:09.467580 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-b9jts" event={"ID":"3e53276a-ccb3-40a1-b1e6-307cc335ff4d","Type":"ContainerStarted","Data":"f62de65d6faaa00abf90ff7f89e431a82fc86dee0f7e8b4eaca5415be527aa05"} Oct 13 08:59:09 crc kubenswrapper[4685]: E1013 08:59:09.477812 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:09deecf840d38ff6af3c924729cf0a9444bc985848bfbe7c918019b88a6bc4d7\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-b9jts" podUID="3e53276a-ccb3-40a1-b1e6-307cc335ff4d" Oct 13 08:59:09 crc kubenswrapper[4685]: I1013 08:59:09.563385 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-bgvk8" podStartSLOduration=4.36774162 podStartE2EDuration="26.563367972s" podCreationTimestamp="2025-10-13 08:58:43 +0000 UTC" firstStartedPulling="2025-10-13 08:58:45.493696991 +0000 UTC m=+850.641572752" lastFinishedPulling="2025-10-13 08:59:07.689323343 +0000 UTC m=+872.837199104" observedRunningTime="2025-10-13 08:59:09.502284778 +0000 UTC m=+874.650160549" watchObservedRunningTime="2025-10-13 08:59:09.563367972 +0000 UTC m=+874.711243723" Oct 13 08:59:09 crc kubenswrapper[4685]: E1013 08:59:09.578396 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-7ktz6" podUID="a4e11db1-f272-469d-9c22-a649cdbcf95e" Oct 13 08:59:10 crc kubenswrapper[4685]: I1013 08:59:10.473980 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-7ktz6" event={"ID":"a4e11db1-f272-469d-9c22-a649cdbcf95e","Type":"ContainerStarted","Data":"d5aba6af0afeab9f4d197ba94f3bd7c5f3444e6eaf9fb0b87802b81d56846db6"} Oct 13 08:59:10 crc kubenswrapper[4685]: I1013 08:59:10.475598 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-xf9rd" event={"ID":"9899b0d7-81a4-49f7-91cd-3c5aa72d49b2","Type":"ContainerStarted","Data":"35c262cd354b154bffb24910538730acdb62ce9e7d512c54bff00e0dac0bee9d"} Oct 13 08:59:10 crc kubenswrapper[4685]: E1013 08:59:10.475745 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-7ktz6" podUID="a4e11db1-f272-469d-9c22-a649cdbcf95e" Oct 13 08:59:10 crc kubenswrapper[4685]: I1013 08:59:10.477113 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-ft77l" event={"ID":"d363a932-2a5d-4082-841a-b5d677d0b1a1","Type":"ContainerStarted","Data":"96d6c3a4503347a25ad67995dc25e031c6319cee8d1f419fe20410c3a88bb70b"} Oct 13 08:59:10 crc kubenswrapper[4685]: I1013 08:59:10.479356 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-fhw55" event={"ID":"9239cb03-ddca-4542-b188-b89717a00f75","Type":"ContainerStarted","Data":"ccecc88e228223fbbd26a2ab16c356fac849e7346b1bd24b6ddfe9f08d8ae553"} Oct 13 08:59:10 crc kubenswrapper[4685]: I1013 08:59:10.486180 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-mj7x8" event={"ID":"6b61f732-3f39-459e-bec0-a6bdd5added4","Type":"ContainerStarted","Data":"f65320751eefa7a41d263ea1be69653a335d6f6b32cf096062ee4233ea53a6fd"} Oct 13 08:59:10 crc kubenswrapper[4685]: E1013 08:59:10.490075 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:783f711b4cb179819cfcb81167c3591c70671440f4551bbe48b7a8730567f577\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-5d7zf" podUID="e509e801-67de-4a55-bd22-cf3f73deca81" Oct 13 08:59:10 crc kubenswrapper[4685]: E1013 08:59:10.490152 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:063a7e65b4ba98f0506f269ff7525b446eae06a5ed4a61c18ffa33a886500867\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-sl5lv" podUID="eb08316b-7fd3-4d65-88e1-bbc91efcb7c7" Oct 13 08:59:10 crc kubenswrapper[4685]: E1013 08:59:10.490221 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:7e584b1c430441c8b6591dadeff32e065de8a185ad37ef90d2e08d37e59aab4a\\\"\"" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-dd92k" podUID="0b4f3fea-3a89-4d28-89c0-436da959e36f" Oct 13 08:59:10 crc kubenswrapper[4685]: E1013 08:59:10.490267 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-cbwvq" podUID="de8c01ad-9e44-42f8-956c-0bb61165a222" Oct 13 08:59:10 crc kubenswrapper[4685]: E1013 08:59:10.490272 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:09deecf840d38ff6af3c924729cf0a9444bc985848bfbe7c918019b88a6bc4d7\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-b9jts" podUID="3e53276a-ccb3-40a1-b1e6-307cc335ff4d" Oct 13 08:59:10 crc kubenswrapper[4685]: E1013 08:59:10.490305 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.174:5001/openstack-k8s-operators/keystone-operator:5d735e29a58b1e4afb30ec827263ac7cab51ba01\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 08:59:11 crc kubenswrapper[4685]: I1013 08:59:11.493568 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-664664cb68-fhw55" Oct 13 08:59:11 crc kubenswrapper[4685]: E1013 08:59:11.494796 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-7ktz6" podUID="a4e11db1-f272-469d-9c22-a649cdbcf95e" Oct 13 08:59:11 crc kubenswrapper[4685]: I1013 08:59:11.538884 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-664664cb68-fhw55" podStartSLOduration=5.245195166 podStartE2EDuration="28.538852188s" podCreationTimestamp="2025-10-13 08:58:43 +0000 UTC" firstStartedPulling="2025-10-13 08:58:45.512458105 +0000 UTC m=+850.660333866" lastFinishedPulling="2025-10-13 08:59:08.806115127 +0000 UTC m=+873.953990888" observedRunningTime="2025-10-13 08:59:11.533851961 +0000 UTC m=+876.681727732" watchObservedRunningTime="2025-10-13 08:59:11.538852188 +0000 UTC m=+876.686727989" Oct 13 08:59:12 crc kubenswrapper[4685]: E1013 08:59:12.068689 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zm9bd" podUID="54405c7c-61f7-41ea-ae0a-29128b51326c" Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.500958 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-xf9rd" event={"ID":"9899b0d7-81a4-49f7-91cd-3c5aa72d49b2","Type":"ContainerStarted","Data":"a1129689cd973719aca7c5de2bf0ec9d02997439e19bd2560023c6777dcdc9d3"} Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.501482 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-xf9rd" Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.503123 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td" event={"ID":"d4c8f24c-6e84-4931-8edb-504b184ea7b0","Type":"ContainerStarted","Data":"deffbe307cc22a695e30beefa4f843e295de17c79e180ddc9b5f71be60cc74df"} Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.505433 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-mj7x8" event={"ID":"6b61f732-3f39-459e-bec0-a6bdd5added4","Type":"ContainerStarted","Data":"615b81927e2f18a1a4f6be5e08e2904268b443213f1d92e75df095da98f21d51"} Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.505611 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-mj7x8" Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.507403 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-q8dkt" event={"ID":"12dc02ce-76d4-4376-ab6f-b2a75580cc4d","Type":"ContainerStarted","Data":"cd3abf2fa55c0d5e9beb5f6126fb148842025fdd87f76460bd70e3bb021f0cad"} Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.510797 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-hbnzh" event={"ID":"92a042d7-669e-48d2-8d7e-8a8da4fc01eb","Type":"ContainerStarted","Data":"a94d4d9c81f7a5d0be5c380111d7a43d34c013d5ce04a7c4e3caa873bbb5594f"} Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.512272 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-2bcss" event={"ID":"a3e83b7f-5017-44e7-b507-f46a2d3f5488","Type":"ContainerStarted","Data":"4ce3c363f1881291ff64d3bd841c4e576ba3387ca25d554008295ea006a67b14"} Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.512684 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-2bcss" Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.521447 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-ft77l" event={"ID":"d363a932-2a5d-4082-841a-b5d677d0b1a1","Type":"ContainerStarted","Data":"73b6bf0916d6197c23550cb61376c6a0618654b55dcc8ffa206a068ecf78aaf7"} Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.522593 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-ft77l" Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.540833 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-9wdvd" event={"ID":"026b2615-8dc1-4ba7-83d1-1e21f4fa80d2","Type":"ContainerStarted","Data":"3c7b99299b8255418f4ca00d40380ad6722b2429fecc0aeded35965b9ab49464"} Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.541742 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-646675d848-9wdvd" Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.561179 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zm9bd" event={"ID":"54405c7c-61f7-41ea-ae0a-29128b51326c","Type":"ContainerStarted","Data":"7fcd8ebf844a496aa6a7483ebf4f29cfbd54a699b1c3da6d3ae75bf800fd56d2"} Oct 13 08:59:12 crc kubenswrapper[4685]: E1013 08:59:12.569796 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:3cc6bba71197ddf88dd4ba1301542bacbc1fe12e6faab2b69e6960944b3d74a0\\\"\"" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zm9bd" podUID="54405c7c-61f7-41ea-ae0a-29128b51326c" Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.570704 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-j65jt" event={"ID":"e297e3a5-6cd9-4314-b4dc-76ba4c25c6f5","Type":"ContainerStarted","Data":"b4f07b04d1538d9387a739b50904f87caa9b7c58a3987c32acb3fe66de157da5"} Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.595186 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-qbl58" event={"ID":"c92c1fba-b02b-4b6c-9570-b75ee60c5e86","Type":"ContainerStarted","Data":"63a15f552f506bd3e76f005985343c037a74982b9ddb6532d86f14f4b56fb003"} Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.607290 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-d6gn6" event={"ID":"d639bb21-69ae-45c8-8a9c-aac17f57f8dd","Type":"ContainerStarted","Data":"ec0c7460228268af4555a34b022958aa44816f7e096483c8922fe05ffc94a1e5"} Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.608024 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-d6gn6" Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.615069 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qnmxx" event={"ID":"f1085de4-f906-4315-b8b5-a3ee4e7182c1","Type":"ContainerStarted","Data":"37f67ed39738efd3e582c300df22a526b059ce75330e5599020b6dc32e3e497a"} Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.656572 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-xf9rd" podStartSLOduration=7.718195178 podStartE2EDuration="30.656558387s" podCreationTimestamp="2025-10-13 08:58:42 +0000 UTC" firstStartedPulling="2025-10-13 08:58:44.753020791 +0000 UTC m=+849.900896552" lastFinishedPulling="2025-10-13 08:59:07.691384 +0000 UTC m=+872.839259761" observedRunningTime="2025-10-13 08:59:12.632742365 +0000 UTC m=+877.780618126" watchObservedRunningTime="2025-10-13 08:59:12.656558387 +0000 UTC m=+877.804434148" Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.712924 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-mj7x8" podStartSLOduration=8.059111918 podStartE2EDuration="30.71289239s" podCreationTimestamp="2025-10-13 08:58:42 +0000 UTC" firstStartedPulling="2025-10-13 08:58:45.067323341 +0000 UTC m=+850.215199102" lastFinishedPulling="2025-10-13 08:59:07.721103803 +0000 UTC m=+872.868979574" observedRunningTime="2025-10-13 08:59:12.711259366 +0000 UTC m=+877.859135127" watchObservedRunningTime="2025-10-13 08:59:12.71289239 +0000 UTC m=+877.860768151" Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.767037 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-2bcss" podStartSLOduration=7.367179343 podStartE2EDuration="30.767011884s" podCreationTimestamp="2025-10-13 08:58:42 +0000 UTC" firstStartedPulling="2025-10-13 08:58:45.641099119 +0000 UTC m=+850.788974880" lastFinishedPulling="2025-10-13 08:59:09.04093166 +0000 UTC m=+874.188807421" observedRunningTime="2025-10-13 08:59:12.761446311 +0000 UTC m=+877.909322072" watchObservedRunningTime="2025-10-13 08:59:12.767011884 +0000 UTC m=+877.914887645" Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.842070 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-d6gn6" podStartSLOduration=6.315235398 podStartE2EDuration="29.842051378s" podCreationTimestamp="2025-10-13 08:58:43 +0000 UTC" firstStartedPulling="2025-10-13 08:58:45.512694201 +0000 UTC m=+850.660569962" lastFinishedPulling="2025-10-13 08:59:09.039510181 +0000 UTC m=+874.187385942" observedRunningTime="2025-10-13 08:59:12.839558711 +0000 UTC m=+877.987434472" watchObservedRunningTime="2025-10-13 08:59:12.842051378 +0000 UTC m=+877.989927139" Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.877066 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-ft77l" podStartSLOduration=7.640832384 podStartE2EDuration="29.877049968s" podCreationTimestamp="2025-10-13 08:58:43 +0000 UTC" firstStartedPulling="2025-10-13 08:58:45.493443894 +0000 UTC m=+850.641319655" lastFinishedPulling="2025-10-13 08:59:07.729661468 +0000 UTC m=+872.877537239" observedRunningTime="2025-10-13 08:59:12.874998591 +0000 UTC m=+878.022874352" watchObservedRunningTime="2025-10-13 08:59:12.877049968 +0000 UTC m=+878.024925729" Oct 13 08:59:12 crc kubenswrapper[4685]: I1013 08:59:12.903059 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-646675d848-9wdvd" podStartSLOduration=6.528276646 podStartE2EDuration="29.90304477s" podCreationTimestamp="2025-10-13 08:58:43 +0000 UTC" firstStartedPulling="2025-10-13 08:58:45.512903237 +0000 UTC m=+850.660778998" lastFinishedPulling="2025-10-13 08:59:08.887671361 +0000 UTC m=+874.035547122" observedRunningTime="2025-10-13 08:59:12.896872581 +0000 UTC m=+878.044748342" watchObservedRunningTime="2025-10-13 08:59:12.90304477 +0000 UTC m=+878.050920531" Oct 13 08:59:13 crc kubenswrapper[4685]: I1013 08:59:13.622085 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-j65jt" event={"ID":"e297e3a5-6cd9-4314-b4dc-76ba4c25c6f5","Type":"ContainerStarted","Data":"c9b8036f87dfd2c4dc6d6e7b939e7d2e7105c61c552af8dfbaba0744d84d21bf"} Oct 13 08:59:13 crc kubenswrapper[4685]: I1013 08:59:13.622150 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-j65jt" Oct 13 08:59:13 crc kubenswrapper[4685]: I1013 08:59:13.623791 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-qbl58" event={"ID":"c92c1fba-b02b-4b6c-9570-b75ee60c5e86","Type":"ContainerStarted","Data":"88df1d3591a773c003d63064147f7ef755a3c62d201f9550c9f9b0a4f157d9b8"} Oct 13 08:59:13 crc kubenswrapper[4685]: I1013 08:59:13.623884 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-qbl58" Oct 13 08:59:13 crc kubenswrapper[4685]: I1013 08:59:13.625457 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-q8dkt" event={"ID":"12dc02ce-76d4-4376-ab6f-b2a75580cc4d","Type":"ContainerStarted","Data":"f24ebdaa13aba1305e9d01de83a53535ac20a92ea3d1b4455da8824504c6835a"} Oct 13 08:59:13 crc kubenswrapper[4685]: I1013 08:59:13.625544 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-59578bc799-q8dkt" Oct 13 08:59:13 crc kubenswrapper[4685]: I1013 08:59:13.627131 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-hbnzh" event={"ID":"92a042d7-669e-48d2-8d7e-8a8da4fc01eb","Type":"ContainerStarted","Data":"1c6270e9a841d22952c8a3e35a751d8ccdd022bc72f221daf7d9a42004489f47"} Oct 13 08:59:13 crc kubenswrapper[4685]: I1013 08:59:13.627249 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-hbnzh" Oct 13 08:59:13 crc kubenswrapper[4685]: I1013 08:59:13.628501 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qnmxx" event={"ID":"f1085de4-f906-4315-b8b5-a3ee4e7182c1","Type":"ContainerStarted","Data":"caf64002e6795946d71d5e560f446d9975ddcaa80a306d5c1c3bab3b6e54a9ea"} Oct 13 08:59:13 crc kubenswrapper[4685]: I1013 08:59:13.628647 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qnmxx" Oct 13 08:59:13 crc kubenswrapper[4685]: I1013 08:59:13.629992 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td" event={"ID":"d4c8f24c-6e84-4931-8edb-504b184ea7b0","Type":"ContainerStarted","Data":"73def7cbe4f82228bfba023ac032b934b43d15841e3e3b7f932beb9a07f985b0"} Oct 13 08:59:13 crc kubenswrapper[4685]: I1013 08:59:13.630558 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td" Oct 13 08:59:13 crc kubenswrapper[4685]: E1013 08:59:13.631476 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:3cc6bba71197ddf88dd4ba1301542bacbc1fe12e6faab2b69e6960944b3d74a0\\\"\"" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zm9bd" podUID="54405c7c-61f7-41ea-ae0a-29128b51326c" Oct 13 08:59:13 crc kubenswrapper[4685]: I1013 08:59:13.654311 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-j65jt" podStartSLOduration=8.459030196 podStartE2EDuration="30.654290869s" podCreationTimestamp="2025-10-13 08:58:43 +0000 UTC" firstStartedPulling="2025-10-13 08:58:45.494003619 +0000 UTC m=+850.641879380" lastFinishedPulling="2025-10-13 08:59:07.689264292 +0000 UTC m=+872.837140053" observedRunningTime="2025-10-13 08:59:13.649658813 +0000 UTC m=+878.797534574" watchObservedRunningTime="2025-10-13 08:59:13.654290869 +0000 UTC m=+878.802166630" Oct 13 08:59:13 crc kubenswrapper[4685]: I1013 08:59:13.671809 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-qbl58" podStartSLOduration=8.464928485 podStartE2EDuration="31.671789239s" podCreationTimestamp="2025-10-13 08:58:42 +0000 UTC" firstStartedPulling="2025-10-13 08:58:44.482949962 +0000 UTC m=+849.630825723" lastFinishedPulling="2025-10-13 08:59:07.689810716 +0000 UTC m=+872.837686477" observedRunningTime="2025-10-13 08:59:13.667693907 +0000 UTC m=+878.815569678" watchObservedRunningTime="2025-10-13 08:59:13.671789239 +0000 UTC m=+878.819665000" Oct 13 08:59:13 crc kubenswrapper[4685]: I1013 08:59:13.695524 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-hbnzh" podStartSLOduration=8.558294651 podStartE2EDuration="31.695502868s" podCreationTimestamp="2025-10-13 08:58:42 +0000 UTC" firstStartedPulling="2025-10-13 08:58:44.581845681 +0000 UTC m=+849.729721442" lastFinishedPulling="2025-10-13 08:59:07.719053898 +0000 UTC m=+872.866929659" observedRunningTime="2025-10-13 08:59:13.692884087 +0000 UTC m=+878.840759848" watchObservedRunningTime="2025-10-13 08:59:13.695502868 +0000 UTC m=+878.843378629" Oct 13 08:59:13 crc kubenswrapper[4685]: I1013 08:59:13.712823 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qnmxx" podStartSLOduration=8.048576462 podStartE2EDuration="30.712806942s" podCreationTimestamp="2025-10-13 08:58:43 +0000 UTC" firstStartedPulling="2025-10-13 08:58:45.056028261 +0000 UTC m=+850.203904012" lastFinishedPulling="2025-10-13 08:59:07.720258731 +0000 UTC m=+872.868134492" observedRunningTime="2025-10-13 08:59:13.709347248 +0000 UTC m=+878.857223009" watchObservedRunningTime="2025-10-13 08:59:13.712806942 +0000 UTC m=+878.860682713" Oct 13 08:59:13 crc kubenswrapper[4685]: I1013 08:59:13.756963 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-59578bc799-q8dkt" podStartSLOduration=7.789819344 podStartE2EDuration="30.756943902s" podCreationTimestamp="2025-10-13 08:58:43 +0000 UTC" firstStartedPulling="2025-10-13 08:58:44.752698361 +0000 UTC m=+849.900574122" lastFinishedPulling="2025-10-13 08:59:07.719822919 +0000 UTC m=+872.867698680" observedRunningTime="2025-10-13 08:59:13.751253176 +0000 UTC m=+878.899128937" watchObservedRunningTime="2025-10-13 08:59:13.756943902 +0000 UTC m=+878.904819663" Oct 13 08:59:13 crc kubenswrapper[4685]: I1013 08:59:13.781434 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td" podStartSLOduration=8.940047893 podStartE2EDuration="30.781412612s" podCreationTimestamp="2025-10-13 08:58:43 +0000 UTC" firstStartedPulling="2025-10-13 08:58:45.850155105 +0000 UTC m=+850.998030856" lastFinishedPulling="2025-10-13 08:59:07.691519814 +0000 UTC m=+872.839395575" observedRunningTime="2025-10-13 08:59:13.777989539 +0000 UTC m=+878.925865300" watchObservedRunningTime="2025-10-13 08:59:13.781412612 +0000 UTC m=+878.929288373" Oct 13 08:59:22 crc kubenswrapper[4685]: I1013 08:59:22.695272 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-5d7zf" event={"ID":"e509e801-67de-4a55-bd22-cf3f73deca81","Type":"ContainerStarted","Data":"45ad7c3203a4a4d543355f60b62790536d7cf562f34e4993ca42b7b16a39f376"} Oct 13 08:59:22 crc kubenswrapper[4685]: I1013 08:59:22.695993 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-5d7zf" Oct 13 08:59:22 crc kubenswrapper[4685]: I1013 08:59:22.712572 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-5d7zf" podStartSLOduration=2.8211770879999998 podStartE2EDuration="40.712557625s" podCreationTimestamp="2025-10-13 08:58:42 +0000 UTC" firstStartedPulling="2025-10-13 08:58:44.021857451 +0000 UTC m=+849.169733212" lastFinishedPulling="2025-10-13 08:59:21.913237988 +0000 UTC m=+887.061113749" observedRunningTime="2025-10-13 08:59:22.709859791 +0000 UTC m=+887.857735552" watchObservedRunningTime="2025-10-13 08:59:22.712557625 +0000 UTC m=+887.860433386" Oct 13 08:59:23 crc kubenswrapper[4685]: I1013 08:59:23.147394 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-qbl58" Oct 13 08:59:23 crc kubenswrapper[4685]: I1013 08:59:23.484989 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-59578bc799-q8dkt" Oct 13 08:59:23 crc kubenswrapper[4685]: I1013 08:59:23.558656 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-hbnzh" Oct 13 08:59:23 crc kubenswrapper[4685]: I1013 08:59:23.572101 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-mj7x8" Oct 13 08:59:23 crc kubenswrapper[4685]: I1013 08:59:23.669238 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-xf9rd" Oct 13 08:59:23 crc kubenswrapper[4685]: I1013 08:59:23.729585 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-b9jts" event={"ID":"3e53276a-ccb3-40a1-b1e6-307cc335ff4d","Type":"ContainerStarted","Data":"43052f4679f0de435774b8d7fa9d84d63d6aa4e93c8d16c3ae4815477932094a"} Oct 13 08:59:23 crc kubenswrapper[4685]: I1013 08:59:23.730098 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-b9jts" Oct 13 08:59:23 crc kubenswrapper[4685]: I1013 08:59:23.764877 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-7ktz6" event={"ID":"a4e11db1-f272-469d-9c22-a649cdbcf95e","Type":"ContainerStarted","Data":"246fa0fc8812503c0725930d8a4168225ef2ede63854c6323251c01dc3341b95"} Oct 13 08:59:23 crc kubenswrapper[4685]: I1013 08:59:23.765245 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-7ktz6" Oct 13 08:59:23 crc kubenswrapper[4685]: I1013 08:59:23.798120 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-7ktz6" podStartSLOduration=3.274528311 podStartE2EDuration="40.798100242s" podCreationTimestamp="2025-10-13 08:58:43 +0000 UTC" firstStartedPulling="2025-10-13 08:58:45.493142546 +0000 UTC m=+850.641018307" lastFinishedPulling="2025-10-13 08:59:23.016714477 +0000 UTC m=+888.164590238" observedRunningTime="2025-10-13 08:59:23.797155647 +0000 UTC m=+888.945031398" watchObservedRunningTime="2025-10-13 08:59:23.798100242 +0000 UTC m=+888.945976003" Oct 13 08:59:23 crc kubenswrapper[4685]: I1013 08:59:23.799159 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-b9jts" podStartSLOduration=2.891530829 podStartE2EDuration="40.799151592s" podCreationTimestamp="2025-10-13 08:58:43 +0000 UTC" firstStartedPulling="2025-10-13 08:58:45.065387747 +0000 UTC m=+850.213263508" lastFinishedPulling="2025-10-13 08:59:22.97300851 +0000 UTC m=+888.120884271" observedRunningTime="2025-10-13 08:59:23.765459129 +0000 UTC m=+888.913334890" watchObservedRunningTime="2025-10-13 08:59:23.799151592 +0000 UTC m=+888.947027353" Oct 13 08:59:23 crc kubenswrapper[4685]: I1013 08:59:23.841766 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qnmxx" Oct 13 08:59:23 crc kubenswrapper[4685]: I1013 08:59:23.890166 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-j65jt" Oct 13 08:59:23 crc kubenswrapper[4685]: I1013 08:59:23.937154 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-664664cb68-fhw55" Oct 13 08:59:23 crc kubenswrapper[4685]: I1013 08:59:23.962601 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-d6gn6" Oct 13 08:59:23 crc kubenswrapper[4685]: I1013 08:59:23.988445 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-ft77l" Oct 13 08:59:24 crc kubenswrapper[4685]: I1013 08:59:24.144606 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-646675d848-9wdvd" Oct 13 08:59:24 crc kubenswrapper[4685]: I1013 08:59:24.774153 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-cbwvq" event={"ID":"de8c01ad-9e44-42f8-956c-0bb61165a222","Type":"ContainerStarted","Data":"09b09a72dc4408ba149f93a796df2e0204be54cf73fe21db7923331555c2c7ca"} Oct 13 08:59:24 crc kubenswrapper[4685]: I1013 08:59:24.775415 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-cbwvq" Oct 13 08:59:24 crc kubenswrapper[4685]: I1013 08:59:24.803949 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-cbwvq" podStartSLOduration=2.86532501 podStartE2EDuration="41.803904705s" podCreationTimestamp="2025-10-13 08:58:43 +0000 UTC" firstStartedPulling="2025-10-13 08:58:45.067589038 +0000 UTC m=+850.215464799" lastFinishedPulling="2025-10-13 08:59:24.006168733 +0000 UTC m=+889.154044494" observedRunningTime="2025-10-13 08:59:24.801634934 +0000 UTC m=+889.949510695" watchObservedRunningTime="2025-10-13 08:59:24.803904705 +0000 UTC m=+889.951780466" Oct 13 08:59:25 crc kubenswrapper[4685]: I1013 08:59:25.111778 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-2bcss" Oct 13 08:59:25 crc kubenswrapper[4685]: I1013 08:59:25.285499 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td" Oct 13 08:59:25 crc kubenswrapper[4685]: I1013 08:59:25.506716 4685 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 13 08:59:25 crc kubenswrapper[4685]: I1013 08:59:25.782843 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" event={"ID":"b001d17a-1aea-44ba-86c5-ba6b312156c1","Type":"ContainerStarted","Data":"1b0c736e43589c77b3b7ea005ec8f55bd09664ac09cc3b4b417ff1bad387941b"} Oct 13 08:59:25 crc kubenswrapper[4685]: I1013 08:59:25.783059 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 08:59:26 crc kubenswrapper[4685]: I1013 08:59:26.791304 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zm9bd" event={"ID":"54405c7c-61f7-41ea-ae0a-29128b51326c","Type":"ContainerStarted","Data":"2eebc73c46f71770470abc3ccecbf0d2eeb95820c20fac8cf7e22d599a1b4721"} Oct 13 08:59:26 crc kubenswrapper[4685]: I1013 08:59:26.791815 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zm9bd" Oct 13 08:59:26 crc kubenswrapper[4685]: I1013 08:59:26.794337 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-sl5lv" event={"ID":"eb08316b-7fd3-4d65-88e1-bbc91efcb7c7","Type":"ContainerStarted","Data":"9ae3fb2369c41da3e3c60942a36d0779338b507ce34dc226c322c8a630e539fc"} Oct 13 08:59:26 crc kubenswrapper[4685]: I1013 08:59:26.794573 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-sl5lv" Oct 13 08:59:26 crc kubenswrapper[4685]: I1013 08:59:26.796410 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-dd92k" event={"ID":"0b4f3fea-3a89-4d28-89c0-436da959e36f","Type":"ContainerStarted","Data":"856d719aa16cc0f88abdaf20335e02f6d1a3c2f6b4e29401fbf2de375fb6f020"} Oct 13 08:59:26 crc kubenswrapper[4685]: I1013 08:59:26.816092 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podStartSLOduration=3.7937241 podStartE2EDuration="44.816068288s" podCreationTimestamp="2025-10-13 08:58:42 +0000 UTC" firstStartedPulling="2025-10-13 08:58:44.561658688 +0000 UTC m=+849.709534449" lastFinishedPulling="2025-10-13 08:59:25.584002876 +0000 UTC m=+890.731878637" observedRunningTime="2025-10-13 08:59:25.804134316 +0000 UTC m=+890.952010077" watchObservedRunningTime="2025-10-13 08:59:26.816068288 +0000 UTC m=+891.963944059" Oct 13 08:59:26 crc kubenswrapper[4685]: I1013 08:59:26.818725 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zm9bd" podStartSLOduration=2.7474651679999997 podStartE2EDuration="44.81871223s" podCreationTimestamp="2025-10-13 08:58:42 +0000 UTC" firstStartedPulling="2025-10-13 08:58:44.022344314 +0000 UTC m=+849.170220065" lastFinishedPulling="2025-10-13 08:59:26.093591366 +0000 UTC m=+891.241467127" observedRunningTime="2025-10-13 08:59:26.813245161 +0000 UTC m=+891.961120922" watchObservedRunningTime="2025-10-13 08:59:26.81871223 +0000 UTC m=+891.966587991" Oct 13 08:59:26 crc kubenswrapper[4685]: I1013 08:59:26.832082 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-dd92k" podStartSLOduration=3.322724601 podStartE2EDuration="43.832044905s" podCreationTimestamp="2025-10-13 08:58:43 +0000 UTC" firstStartedPulling="2025-10-13 08:58:45.494842702 +0000 UTC m=+850.642718463" lastFinishedPulling="2025-10-13 08:59:26.004162996 +0000 UTC m=+891.152038767" observedRunningTime="2025-10-13 08:59:26.826717009 +0000 UTC m=+891.974592780" watchObservedRunningTime="2025-10-13 08:59:26.832044905 +0000 UTC m=+891.979920676" Oct 13 08:59:26 crc kubenswrapper[4685]: I1013 08:59:26.849972 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-sl5lv" podStartSLOduration=3.532105833 podStartE2EDuration="44.849953646s" podCreationTimestamp="2025-10-13 08:58:42 +0000 UTC" firstStartedPulling="2025-10-13 08:58:44.77269596 +0000 UTC m=+849.920571721" lastFinishedPulling="2025-10-13 08:59:26.090543773 +0000 UTC m=+891.238419534" observedRunningTime="2025-10-13 08:59:26.84716758 +0000 UTC m=+891.995043341" watchObservedRunningTime="2025-10-13 08:59:26.849953646 +0000 UTC m=+891.997829417" Oct 13 08:59:33 crc kubenswrapper[4685]: I1013 08:59:33.141966 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-5d7zf" Oct 13 08:59:33 crc kubenswrapper[4685]: I1013 08:59:33.238401 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zm9bd" Oct 13 08:59:33 crc kubenswrapper[4685]: I1013 08:59:33.444604 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 08:59:33 crc kubenswrapper[4685]: I1013 08:59:33.579510 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-sl5lv" Oct 13 08:59:33 crc kubenswrapper[4685]: I1013 08:59:33.675534 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-cbwvq" Oct 13 08:59:33 crc kubenswrapper[4685]: I1013 08:59:33.748476 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-b9jts" Oct 13 08:59:33 crc kubenswrapper[4685]: I1013 08:59:33.810993 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-7ktz6" Oct 13 08:59:34 crc kubenswrapper[4685]: I1013 08:59:34.090024 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-dd92k" Oct 13 08:59:34 crc kubenswrapper[4685]: I1013 08:59:34.094495 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-dd92k" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.246278 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-bztxj"] Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.247750 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-bztxj" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.260036 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-bjkmv" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.262858 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.264339 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.266011 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.282897 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-bztxj"] Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.385340 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fgdd\" (UniqueName: \"kubernetes.io/projected/e7118b17-2973-4bc3-b5b3-027032ff0119-kube-api-access-7fgdd\") pod \"dnsmasq-dns-675f4bcbfc-bztxj\" (UID: \"e7118b17-2973-4bc3-b5b3-027032ff0119\") " pod="openstack/dnsmasq-dns-675f4bcbfc-bztxj" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.385405 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7118b17-2973-4bc3-b5b3-027032ff0119-config\") pod \"dnsmasq-dns-675f4bcbfc-bztxj\" (UID: \"e7118b17-2973-4bc3-b5b3-027032ff0119\") " pod="openstack/dnsmasq-dns-675f4bcbfc-bztxj" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.421390 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-89xzl"] Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.422797 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-89xzl" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.425210 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.446105 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-89xzl"] Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.486069 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2134acd-f351-4c06-acc6-b2453a8b0864-config\") pod \"dnsmasq-dns-78dd6ddcc-89xzl\" (UID: \"b2134acd-f351-4c06-acc6-b2453a8b0864\") " pod="openstack/dnsmasq-dns-78dd6ddcc-89xzl" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.486368 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fgdd\" (UniqueName: \"kubernetes.io/projected/e7118b17-2973-4bc3-b5b3-027032ff0119-kube-api-access-7fgdd\") pod \"dnsmasq-dns-675f4bcbfc-bztxj\" (UID: \"e7118b17-2973-4bc3-b5b3-027032ff0119\") " pod="openstack/dnsmasq-dns-675f4bcbfc-bztxj" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.486474 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2134acd-f351-4c06-acc6-b2453a8b0864-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-89xzl\" (UID: \"b2134acd-f351-4c06-acc6-b2453a8b0864\") " pod="openstack/dnsmasq-dns-78dd6ddcc-89xzl" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.486618 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7118b17-2973-4bc3-b5b3-027032ff0119-config\") pod \"dnsmasq-dns-675f4bcbfc-bztxj\" (UID: \"e7118b17-2973-4bc3-b5b3-027032ff0119\") " pod="openstack/dnsmasq-dns-675f4bcbfc-bztxj" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.486659 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdj26\" (UniqueName: \"kubernetes.io/projected/b2134acd-f351-4c06-acc6-b2453a8b0864-kube-api-access-sdj26\") pod \"dnsmasq-dns-78dd6ddcc-89xzl\" (UID: \"b2134acd-f351-4c06-acc6-b2453a8b0864\") " pod="openstack/dnsmasq-dns-78dd6ddcc-89xzl" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.487713 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7118b17-2973-4bc3-b5b3-027032ff0119-config\") pod \"dnsmasq-dns-675f4bcbfc-bztxj\" (UID: \"e7118b17-2973-4bc3-b5b3-027032ff0119\") " pod="openstack/dnsmasq-dns-675f4bcbfc-bztxj" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.520974 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fgdd\" (UniqueName: \"kubernetes.io/projected/e7118b17-2973-4bc3-b5b3-027032ff0119-kube-api-access-7fgdd\") pod \"dnsmasq-dns-675f4bcbfc-bztxj\" (UID: \"e7118b17-2973-4bc3-b5b3-027032ff0119\") " pod="openstack/dnsmasq-dns-675f4bcbfc-bztxj" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.562094 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-bztxj" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.596533 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2134acd-f351-4c06-acc6-b2453a8b0864-config\") pod \"dnsmasq-dns-78dd6ddcc-89xzl\" (UID: \"b2134acd-f351-4c06-acc6-b2453a8b0864\") " pod="openstack/dnsmasq-dns-78dd6ddcc-89xzl" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.596621 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2134acd-f351-4c06-acc6-b2453a8b0864-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-89xzl\" (UID: \"b2134acd-f351-4c06-acc6-b2453a8b0864\") " pod="openstack/dnsmasq-dns-78dd6ddcc-89xzl" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.596654 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdj26\" (UniqueName: \"kubernetes.io/projected/b2134acd-f351-4c06-acc6-b2453a8b0864-kube-api-access-sdj26\") pod \"dnsmasq-dns-78dd6ddcc-89xzl\" (UID: \"b2134acd-f351-4c06-acc6-b2453a8b0864\") " pod="openstack/dnsmasq-dns-78dd6ddcc-89xzl" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.597565 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2134acd-f351-4c06-acc6-b2453a8b0864-config\") pod \"dnsmasq-dns-78dd6ddcc-89xzl\" (UID: \"b2134acd-f351-4c06-acc6-b2453a8b0864\") " pod="openstack/dnsmasq-dns-78dd6ddcc-89xzl" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.598262 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2134acd-f351-4c06-acc6-b2453a8b0864-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-89xzl\" (UID: \"b2134acd-f351-4c06-acc6-b2453a8b0864\") " pod="openstack/dnsmasq-dns-78dd6ddcc-89xzl" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.622721 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdj26\" (UniqueName: \"kubernetes.io/projected/b2134acd-f351-4c06-acc6-b2453a8b0864-kube-api-access-sdj26\") pod \"dnsmasq-dns-78dd6ddcc-89xzl\" (UID: \"b2134acd-f351-4c06-acc6-b2453a8b0864\") " pod="openstack/dnsmasq-dns-78dd6ddcc-89xzl" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.735643 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-89xzl" Oct 13 08:59:50 crc kubenswrapper[4685]: I1013 08:59:50.981789 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-89xzl"] Oct 13 08:59:50 crc kubenswrapper[4685]: W1013 08:59:50.986939 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2134acd_f351_4c06_acc6_b2453a8b0864.slice/crio-ca6f5cfa32ec99b7c7675bdc33362168a626c01ef78603de30041fa488a8cda6 WatchSource:0}: Error finding container ca6f5cfa32ec99b7c7675bdc33362168a626c01ef78603de30041fa488a8cda6: Status 404 returned error can't find the container with id ca6f5cfa32ec99b7c7675bdc33362168a626c01ef78603de30041fa488a8cda6 Oct 13 08:59:51 crc kubenswrapper[4685]: W1013 08:59:51.017480 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7118b17_2973_4bc3_b5b3_027032ff0119.slice/crio-2c050d6e047cbaa869699ca64fa413d046f73bf8a91ec55f2b5b8b227bd8a035 WatchSource:0}: Error finding container 2c050d6e047cbaa869699ca64fa413d046f73bf8a91ec55f2b5b8b227bd8a035: Status 404 returned error can't find the container with id 2c050d6e047cbaa869699ca64fa413d046f73bf8a91ec55f2b5b8b227bd8a035 Oct 13 08:59:51 crc kubenswrapper[4685]: I1013 08:59:51.017723 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-bztxj"] Oct 13 08:59:51 crc kubenswrapper[4685]: I1013 08:59:51.027145 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-89xzl" event={"ID":"b2134acd-f351-4c06-acc6-b2453a8b0864","Type":"ContainerStarted","Data":"ca6f5cfa32ec99b7c7675bdc33362168a626c01ef78603de30041fa488a8cda6"} Oct 13 08:59:52 crc kubenswrapper[4685]: I1013 08:59:52.035773 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-bztxj" event={"ID":"e7118b17-2973-4bc3-b5b3-027032ff0119","Type":"ContainerStarted","Data":"2c050d6e047cbaa869699ca64fa413d046f73bf8a91ec55f2b5b8b227bd8a035"} Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.118865 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-bztxj"] Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.148114 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-rcf7p"] Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.149266 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-rcf7p" Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.155430 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-rcf7p"] Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.339525 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5-config\") pod \"dnsmasq-dns-666b6646f7-rcf7p\" (UID: \"a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5\") " pod="openstack/dnsmasq-dns-666b6646f7-rcf7p" Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.339603 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx522\" (UniqueName: \"kubernetes.io/projected/a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5-kube-api-access-gx522\") pod \"dnsmasq-dns-666b6646f7-rcf7p\" (UID: \"a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5\") " pod="openstack/dnsmasq-dns-666b6646f7-rcf7p" Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.339716 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5-dns-svc\") pod \"dnsmasq-dns-666b6646f7-rcf7p\" (UID: \"a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5\") " pod="openstack/dnsmasq-dns-666b6646f7-rcf7p" Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.411808 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-89xzl"] Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.435594 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-znf4k"] Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.436864 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-znf4k" Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.440690 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5-config\") pod \"dnsmasq-dns-666b6646f7-rcf7p\" (UID: \"a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5\") " pod="openstack/dnsmasq-dns-666b6646f7-rcf7p" Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.442266 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5-config\") pod \"dnsmasq-dns-666b6646f7-rcf7p\" (UID: \"a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5\") " pod="openstack/dnsmasq-dns-666b6646f7-rcf7p" Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.443007 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx522\" (UniqueName: \"kubernetes.io/projected/a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5-kube-api-access-gx522\") pod \"dnsmasq-dns-666b6646f7-rcf7p\" (UID: \"a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5\") " pod="openstack/dnsmasq-dns-666b6646f7-rcf7p" Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.443157 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5-dns-svc\") pod \"dnsmasq-dns-666b6646f7-rcf7p\" (UID: \"a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5\") " pod="openstack/dnsmasq-dns-666b6646f7-rcf7p" Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.444116 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5-dns-svc\") pod \"dnsmasq-dns-666b6646f7-rcf7p\" (UID: \"a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5\") " pod="openstack/dnsmasq-dns-666b6646f7-rcf7p" Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.452740 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-znf4k"] Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.479716 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx522\" (UniqueName: \"kubernetes.io/projected/a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5-kube-api-access-gx522\") pod \"dnsmasq-dns-666b6646f7-rcf7p\" (UID: \"a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5\") " pod="openstack/dnsmasq-dns-666b6646f7-rcf7p" Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.547023 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-znf4k\" (UID: \"9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc\") " pod="openstack/dnsmasq-dns-57d769cc4f-znf4k" Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.547071 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc-config\") pod \"dnsmasq-dns-57d769cc4f-znf4k\" (UID: \"9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc\") " pod="openstack/dnsmasq-dns-57d769cc4f-znf4k" Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.547603 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hl75\" (UniqueName: \"kubernetes.io/projected/9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc-kube-api-access-5hl75\") pod \"dnsmasq-dns-57d769cc4f-znf4k\" (UID: \"9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc\") " pod="openstack/dnsmasq-dns-57d769cc4f-znf4k" Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.648778 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-znf4k\" (UID: \"9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc\") " pod="openstack/dnsmasq-dns-57d769cc4f-znf4k" Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.648834 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc-config\") pod \"dnsmasq-dns-57d769cc4f-znf4k\" (UID: \"9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc\") " pod="openstack/dnsmasq-dns-57d769cc4f-znf4k" Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.648874 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hl75\" (UniqueName: \"kubernetes.io/projected/9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc-kube-api-access-5hl75\") pod \"dnsmasq-dns-57d769cc4f-znf4k\" (UID: \"9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc\") " pod="openstack/dnsmasq-dns-57d769cc4f-znf4k" Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.650877 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc-config\") pod \"dnsmasq-dns-57d769cc4f-znf4k\" (UID: \"9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc\") " pod="openstack/dnsmasq-dns-57d769cc4f-znf4k" Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.654525 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-znf4k\" (UID: \"9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc\") " pod="openstack/dnsmasq-dns-57d769cc4f-znf4k" Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.668989 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hl75\" (UniqueName: \"kubernetes.io/projected/9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc-kube-api-access-5hl75\") pod \"dnsmasq-dns-57d769cc4f-znf4k\" (UID: \"9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc\") " pod="openstack/dnsmasq-dns-57d769cc4f-znf4k" Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.774007 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-znf4k" Oct 13 08:59:53 crc kubenswrapper[4685]: I1013 08:59:53.774465 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-rcf7p" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.287681 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.289295 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.295331 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-b9zgt" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.295536 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.296365 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.296923 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.296999 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.297161 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.297339 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.306174 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.338438 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-znf4k"] Oct 13 08:59:54 crc kubenswrapper[4685]: W1013 08:59:54.349377 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9df34c78_f8c7_4b0f_9f74_47e1fe50a3fc.slice/crio-8609e32c6cf331898dc5abb70a4fc99f765a56eab90d257388f9c51139de47bc WatchSource:0}: Error finding container 8609e32c6cf331898dc5abb70a4fc99f765a56eab90d257388f9c51139de47bc: Status 404 returned error can't find the container with id 8609e32c6cf331898dc5abb70a4fc99f765a56eab90d257388f9c51139de47bc Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.407898 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-rcf7p"] Oct 13 08:59:54 crc kubenswrapper[4685]: W1013 08:59:54.413882 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda54ac13e_48cb_4c0b_8cf9_e7f01e48b8e5.slice/crio-804c70c18f0b8a9e049ccf3b0e9618018a149c9e039412d35dc0ae91a2fa1a4d WatchSource:0}: Error finding container 804c70c18f0b8a9e049ccf3b0e9618018a149c9e039412d35dc0ae91a2fa1a4d: Status 404 returned error can't find the container with id 804c70c18f0b8a9e049ccf3b0e9618018a149c9e039412d35dc0ae91a2fa1a4d Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.461580 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3b0348a6-6814-47f3-919e-e975c01e00e8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.461651 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.461696 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.461746 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.461772 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3b0348a6-6814-47f3-919e-e975c01e00e8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.461792 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3b0348a6-6814-47f3-919e-e975c01e00e8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.461833 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.461852 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3b0348a6-6814-47f3-919e-e975c01e00e8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.461871 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4kzp\" (UniqueName: \"kubernetes.io/projected/3b0348a6-6814-47f3-919e-e975c01e00e8-kube-api-access-v4kzp\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.461959 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b0348a6-6814-47f3-919e-e975c01e00e8-config-data\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.461992 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.563761 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b0348a6-6814-47f3-919e-e975c01e00e8-config-data\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.563894 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.563967 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3b0348a6-6814-47f3-919e-e975c01e00e8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.564030 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.564058 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.564112 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.564148 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3b0348a6-6814-47f3-919e-e975c01e00e8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.564214 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3b0348a6-6814-47f3-919e-e975c01e00e8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.564271 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.564296 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3b0348a6-6814-47f3-919e-e975c01e00e8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.564320 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4kzp\" (UniqueName: \"kubernetes.io/projected/3b0348a6-6814-47f3-919e-e975c01e00e8-kube-api-access-v4kzp\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.566373 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b0348a6-6814-47f3-919e-e975c01e00e8-config-data\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.566392 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.566412 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.566898 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.567267 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3b0348a6-6814-47f3-919e-e975c01e00e8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.572364 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3b0348a6-6814-47f3-919e-e975c01e00e8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.573899 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.574480 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3b0348a6-6814-47f3-919e-e975c01e00e8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.578229 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.586812 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3b0348a6-6814-47f3-919e-e975c01e00e8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.590942 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.620007 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4kzp\" (UniqueName: \"kubernetes.io/projected/3b0348a6-6814-47f3-919e-e975c01e00e8-kube-api-access-v4kzp\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.644412 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.646083 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.650685 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.650874 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.651902 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-5c4fv" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.652105 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.652247 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.652837 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.653257 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.669428 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.775844 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.776218 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.776253 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b7f19239-a2d9-49b5-80b0-43543daf1787-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.776279 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.776320 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b7f19239-a2d9-49b5-80b0-43543daf1787-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.776340 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b7f19239-a2d9-49b5-80b0-43543daf1787-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.776357 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b7f19239-a2d9-49b5-80b0-43543daf1787-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.776374 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.776395 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b7f19239-a2d9-49b5-80b0-43543daf1787-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.776693 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.776760 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc47f\" (UniqueName: \"kubernetes.io/projected/b7f19239-a2d9-49b5-80b0-43543daf1787-kube-api-access-sc47f\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.878872 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.878946 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc47f\" (UniqueName: \"kubernetes.io/projected/b7f19239-a2d9-49b5-80b0-43543daf1787-kube-api-access-sc47f\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.879008 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.879041 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.879081 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b7f19239-a2d9-49b5-80b0-43543daf1787-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.879111 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.879141 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b7f19239-a2d9-49b5-80b0-43543daf1787-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.879163 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b7f19239-a2d9-49b5-80b0-43543daf1787-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.879187 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b7f19239-a2d9-49b5-80b0-43543daf1787-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.879210 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.879241 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b7f19239-a2d9-49b5-80b0-43543daf1787-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.879840 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.880363 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.880625 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b7f19239-a2d9-49b5-80b0-43543daf1787-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.880650 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b7f19239-a2d9-49b5-80b0-43543daf1787-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.880728 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b7f19239-a2d9-49b5-80b0-43543daf1787-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.880893 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.891533 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.894294 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.898783 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b7f19239-a2d9-49b5-80b0-43543daf1787-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.926676 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b7f19239-a2d9-49b5-80b0-43543daf1787-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.931852 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.933060 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 13 08:59:54 crc kubenswrapper[4685]: I1013 08:59:54.937667 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc47f\" (UniqueName: \"kubernetes.io/projected/b7f19239-a2d9-49b5-80b0-43543daf1787-kube-api-access-sc47f\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:55 crc kubenswrapper[4685]: I1013 08:59:55.003095 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 13 08:59:55 crc kubenswrapper[4685]: I1013 08:59:55.099363 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-rcf7p" event={"ID":"a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5","Type":"ContainerStarted","Data":"804c70c18f0b8a9e049ccf3b0e9618018a149c9e039412d35dc0ae91a2fa1a4d"} Oct 13 08:59:55 crc kubenswrapper[4685]: I1013 08:59:55.117516 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-znf4k" event={"ID":"9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc","Type":"ContainerStarted","Data":"8609e32c6cf331898dc5abb70a4fc99f765a56eab90d257388f9c51139de47bc"} Oct 13 08:59:55 crc kubenswrapper[4685]: I1013 08:59:55.618617 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 13 08:59:55 crc kubenswrapper[4685]: I1013 08:59:55.672712 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.073076 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.074460 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.084990 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.085247 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.085867 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.086017 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.086213 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-729kn" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.089383 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.092597 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.137747 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b7f19239-a2d9-49b5-80b0-43543daf1787","Type":"ContainerStarted","Data":"4355a58fb21aefb60a8f8b4cff96124de1105ec1603c270f7504162362dcbf32"} Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.138610 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3b0348a6-6814-47f3-919e-e975c01e00e8","Type":"ContainerStarted","Data":"42c24f17559cb28c52a00ab2cc29b06472d9e601bebc2c8573f8bb35d9a68633"} Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.208442 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f5482688-fe73-4196-ae4d-a306a1c31545-secrets\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.208479 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkr28\" (UniqueName: \"kubernetes.io/projected/f5482688-fe73-4196-ae4d-a306a1c31545-kube-api-access-nkr28\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.208880 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5482688-fe73-4196-ae4d-a306a1c31545-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.208934 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5482688-fe73-4196-ae4d-a306a1c31545-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.208958 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f5482688-fe73-4196-ae4d-a306a1c31545-kolla-config\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.208973 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f5482688-fe73-4196-ae4d-a306a1c31545-config-data-default\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.209010 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.209026 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f5482688-fe73-4196-ae4d-a306a1c31545-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.209041 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5482688-fe73-4196-ae4d-a306a1c31545-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.313441 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f5482688-fe73-4196-ae4d-a306a1c31545-secrets\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.313511 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkr28\" (UniqueName: \"kubernetes.io/projected/f5482688-fe73-4196-ae4d-a306a1c31545-kube-api-access-nkr28\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.313534 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5482688-fe73-4196-ae4d-a306a1c31545-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.313562 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5482688-fe73-4196-ae4d-a306a1c31545-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.313591 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f5482688-fe73-4196-ae4d-a306a1c31545-kolla-config\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.313610 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f5482688-fe73-4196-ae4d-a306a1c31545-config-data-default\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.313655 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.313680 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f5482688-fe73-4196-ae4d-a306a1c31545-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.313698 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5482688-fe73-4196-ae4d-a306a1c31545-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.315695 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f5482688-fe73-4196-ae4d-a306a1c31545-config-data-default\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.316543 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.317711 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f5482688-fe73-4196-ae4d-a306a1c31545-kolla-config\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.318088 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f5482688-fe73-4196-ae4d-a306a1c31545-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.323839 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5482688-fe73-4196-ae4d-a306a1c31545-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.333713 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f5482688-fe73-4196-ae4d-a306a1c31545-secrets\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.338603 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5482688-fe73-4196-ae4d-a306a1c31545-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.343051 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5482688-fe73-4196-ae4d-a306a1c31545-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.349506 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkr28\" (UniqueName: \"kubernetes.io/projected/f5482688-fe73-4196-ae4d-a306a1c31545-kube-api-access-nkr28\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.372245 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"f5482688-fe73-4196-ae4d-a306a1c31545\") " pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.398292 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 13 08:59:56 crc kubenswrapper[4685]: I1013 08:59:56.838774 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.345572 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.352181 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.352287 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.368650 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-clkbk" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.368770 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.369028 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.370115 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.448032 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.448092 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.448121 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6mht\" (UniqueName: \"kubernetes.io/projected/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-kube-api-access-s6mht\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.448174 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.448194 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.448216 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.448231 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.448249 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.448280 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.458132 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.459315 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.466732 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.466878 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.468869 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-8j4ch" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.476817 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.549629 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5d643a-96c6-4fca-904d-3b4a71e9630c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fb5d643a-96c6-4fca-904d-3b4a71e9630c\") " pod="openstack/memcached-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.549697 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.549720 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.549777 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.549794 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.549834 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fb5d643a-96c6-4fca-904d-3b4a71e9630c-kolla-config\") pod \"memcached-0\" (UID: \"fb5d643a-96c6-4fca-904d-3b4a71e9630c\") " pod="openstack/memcached-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.549853 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.549937 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.549959 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwp4z\" (UniqueName: \"kubernetes.io/projected/fb5d643a-96c6-4fca-904d-3b4a71e9630c-kube-api-access-cwp4z\") pod \"memcached-0\" (UID: \"fb5d643a-96c6-4fca-904d-3b4a71e9630c\") " pod="openstack/memcached-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.549998 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.550026 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.550071 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6mht\" (UniqueName: \"kubernetes.io/projected/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-kube-api-access-s6mht\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.550100 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5d643a-96c6-4fca-904d-3b4a71e9630c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fb5d643a-96c6-4fca-904d-3b4a71e9630c\") " pod="openstack/memcached-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.550117 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb5d643a-96c6-4fca-904d-3b4a71e9630c-config-data\") pod \"memcached-0\" (UID: \"fb5d643a-96c6-4fca-904d-3b4a71e9630c\") " pod="openstack/memcached-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.551310 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.551612 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.551801 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.553112 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.553470 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.564770 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.570959 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.591745 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.605990 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6mht\" (UniqueName: \"kubernetes.io/projected/66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20-kube-api-access-s6mht\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.606395 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20\") " pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.651284 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5d643a-96c6-4fca-904d-3b4a71e9630c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fb5d643a-96c6-4fca-904d-3b4a71e9630c\") " pod="openstack/memcached-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.651448 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fb5d643a-96c6-4fca-904d-3b4a71e9630c-kolla-config\") pod \"memcached-0\" (UID: \"fb5d643a-96c6-4fca-904d-3b4a71e9630c\") " pod="openstack/memcached-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.651511 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwp4z\" (UniqueName: \"kubernetes.io/projected/fb5d643a-96c6-4fca-904d-3b4a71e9630c-kube-api-access-cwp4z\") pod \"memcached-0\" (UID: \"fb5d643a-96c6-4fca-904d-3b4a71e9630c\") " pod="openstack/memcached-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.651606 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5d643a-96c6-4fca-904d-3b4a71e9630c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fb5d643a-96c6-4fca-904d-3b4a71e9630c\") " pod="openstack/memcached-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.651645 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb5d643a-96c6-4fca-904d-3b4a71e9630c-config-data\") pod \"memcached-0\" (UID: \"fb5d643a-96c6-4fca-904d-3b4a71e9630c\") " pod="openstack/memcached-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.657609 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fb5d643a-96c6-4fca-904d-3b4a71e9630c-kolla-config\") pod \"memcached-0\" (UID: \"fb5d643a-96c6-4fca-904d-3b4a71e9630c\") " pod="openstack/memcached-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.663399 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb5d643a-96c6-4fca-904d-3b4a71e9630c-config-data\") pod \"memcached-0\" (UID: \"fb5d643a-96c6-4fca-904d-3b4a71e9630c\") " pod="openstack/memcached-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.676077 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5d643a-96c6-4fca-904d-3b4a71e9630c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fb5d643a-96c6-4fca-904d-3b4a71e9630c\") " pod="openstack/memcached-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.677735 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5d643a-96c6-4fca-904d-3b4a71e9630c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fb5d643a-96c6-4fca-904d-3b4a71e9630c\") " pod="openstack/memcached-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.688699 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.694439 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwp4z\" (UniqueName: \"kubernetes.io/projected/fb5d643a-96c6-4fca-904d-3b4a71e9630c-kube-api-access-cwp4z\") pod \"memcached-0\" (UID: \"fb5d643a-96c6-4fca-904d-3b4a71e9630c\") " pod="openstack/memcached-0" Oct 13 08:59:57 crc kubenswrapper[4685]: I1013 08:59:57.960459 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 13 08:59:59 crc kubenswrapper[4685]: I1013 08:59:59.303274 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 13 08:59:59 crc kubenswrapper[4685]: I1013 08:59:59.304406 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 13 08:59:59 crc kubenswrapper[4685]: I1013 08:59:59.308438 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-7jdwc" Oct 13 08:59:59 crc kubenswrapper[4685]: I1013 08:59:59.378605 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 13 08:59:59 crc kubenswrapper[4685]: I1013 08:59:59.410573 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q29x\" (UniqueName: \"kubernetes.io/projected/eb9b10e0-0158-4c15-9de6-f8f71ecf531b-kube-api-access-7q29x\") pod \"kube-state-metrics-0\" (UID: \"eb9b10e0-0158-4c15-9de6-f8f71ecf531b\") " pod="openstack/kube-state-metrics-0" Oct 13 08:59:59 crc kubenswrapper[4685]: I1013 08:59:59.511594 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q29x\" (UniqueName: \"kubernetes.io/projected/eb9b10e0-0158-4c15-9de6-f8f71ecf531b-kube-api-access-7q29x\") pod \"kube-state-metrics-0\" (UID: \"eb9b10e0-0158-4c15-9de6-f8f71ecf531b\") " pod="openstack/kube-state-metrics-0" Oct 13 08:59:59 crc kubenswrapper[4685]: I1013 08:59:59.538567 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q29x\" (UniqueName: \"kubernetes.io/projected/eb9b10e0-0158-4c15-9de6-f8f71ecf531b-kube-api-access-7q29x\") pod \"kube-state-metrics-0\" (UID: \"eb9b10e0-0158-4c15-9de6-f8f71ecf531b\") " pod="openstack/kube-state-metrics-0" Oct 13 08:59:59 crc kubenswrapper[4685]: I1013 08:59:59.630554 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 13 09:00:00 crc kubenswrapper[4685]: I1013 09:00:00.132894 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339100-h77rw"] Oct 13 09:00:00 crc kubenswrapper[4685]: I1013 09:00:00.133898 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339100-h77rw" Oct 13 09:00:00 crc kubenswrapper[4685]: I1013 09:00:00.135822 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 13 09:00:00 crc kubenswrapper[4685]: I1013 09:00:00.143257 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 13 09:00:00 crc kubenswrapper[4685]: I1013 09:00:00.149506 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339100-h77rw"] Oct 13 09:00:00 crc kubenswrapper[4685]: I1013 09:00:00.221775 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/edbf1439-6678-4cef-a3ec-ba71bbe1bba6-secret-volume\") pod \"collect-profiles-29339100-h77rw\" (UID: \"edbf1439-6678-4cef-a3ec-ba71bbe1bba6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339100-h77rw" Oct 13 09:00:00 crc kubenswrapper[4685]: I1013 09:00:00.221831 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/edbf1439-6678-4cef-a3ec-ba71bbe1bba6-config-volume\") pod \"collect-profiles-29339100-h77rw\" (UID: \"edbf1439-6678-4cef-a3ec-ba71bbe1bba6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339100-h77rw" Oct 13 09:00:00 crc kubenswrapper[4685]: I1013 09:00:00.221865 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n68ss\" (UniqueName: \"kubernetes.io/projected/edbf1439-6678-4cef-a3ec-ba71bbe1bba6-kube-api-access-n68ss\") pod \"collect-profiles-29339100-h77rw\" (UID: \"edbf1439-6678-4cef-a3ec-ba71bbe1bba6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339100-h77rw" Oct 13 09:00:00 crc kubenswrapper[4685]: I1013 09:00:00.322658 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/edbf1439-6678-4cef-a3ec-ba71bbe1bba6-secret-volume\") pod \"collect-profiles-29339100-h77rw\" (UID: \"edbf1439-6678-4cef-a3ec-ba71bbe1bba6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339100-h77rw" Oct 13 09:00:00 crc kubenswrapper[4685]: I1013 09:00:00.322760 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/edbf1439-6678-4cef-a3ec-ba71bbe1bba6-config-volume\") pod \"collect-profiles-29339100-h77rw\" (UID: \"edbf1439-6678-4cef-a3ec-ba71bbe1bba6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339100-h77rw" Oct 13 09:00:00 crc kubenswrapper[4685]: I1013 09:00:00.322801 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n68ss\" (UniqueName: \"kubernetes.io/projected/edbf1439-6678-4cef-a3ec-ba71bbe1bba6-kube-api-access-n68ss\") pod \"collect-profiles-29339100-h77rw\" (UID: \"edbf1439-6678-4cef-a3ec-ba71bbe1bba6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339100-h77rw" Oct 13 09:00:00 crc kubenswrapper[4685]: I1013 09:00:00.324735 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/edbf1439-6678-4cef-a3ec-ba71bbe1bba6-config-volume\") pod \"collect-profiles-29339100-h77rw\" (UID: \"edbf1439-6678-4cef-a3ec-ba71bbe1bba6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339100-h77rw" Oct 13 09:00:00 crc kubenswrapper[4685]: I1013 09:00:00.338358 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n68ss\" (UniqueName: \"kubernetes.io/projected/edbf1439-6678-4cef-a3ec-ba71bbe1bba6-kube-api-access-n68ss\") pod \"collect-profiles-29339100-h77rw\" (UID: \"edbf1439-6678-4cef-a3ec-ba71bbe1bba6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339100-h77rw" Oct 13 09:00:00 crc kubenswrapper[4685]: I1013 09:00:00.343613 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/edbf1439-6678-4cef-a3ec-ba71bbe1bba6-secret-volume\") pod \"collect-profiles-29339100-h77rw\" (UID: \"edbf1439-6678-4cef-a3ec-ba71bbe1bba6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339100-h77rw" Oct 13 09:00:00 crc kubenswrapper[4685]: I1013 09:00:00.461861 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339100-h77rw" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.287605 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f5482688-fe73-4196-ae4d-a306a1c31545","Type":"ContainerStarted","Data":"558a987cec5f15e4cbbed4c3af02ff43be7038ca6b7010f7f777905ca98afaab"} Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.469713 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-tb2c4"] Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.471028 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.477489 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-gh7ls"] Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.479068 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.479651 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.479968 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-6vndn" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.480430 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.487264 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tb2c4"] Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.568329 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-gh7ls"] Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.591215 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9dccff31-26fc-43c7-a679-01173d2f34a9-var-log-ovn\") pod \"ovn-controller-tb2c4\" (UID: \"9dccff31-26fc-43c7-a679-01173d2f34a9\") " pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.591262 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dccff31-26fc-43c7-a679-01173d2f34a9-ovn-controller-tls-certs\") pod \"ovn-controller-tb2c4\" (UID: \"9dccff31-26fc-43c7-a679-01173d2f34a9\") " pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.591306 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dccff31-26fc-43c7-a679-01173d2f34a9-combined-ca-bundle\") pod \"ovn-controller-tb2c4\" (UID: \"9dccff31-26fc-43c7-a679-01173d2f34a9\") " pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.591331 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fb7ad71b-d500-457d-81d4-831ba362a8b4-etc-ovs\") pod \"ovn-controller-ovs-gh7ls\" (UID: \"fb7ad71b-d500-457d-81d4-831ba362a8b4\") " pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.591369 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9dccff31-26fc-43c7-a679-01173d2f34a9-var-run\") pod \"ovn-controller-tb2c4\" (UID: \"9dccff31-26fc-43c7-a679-01173d2f34a9\") " pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.591397 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fb7ad71b-d500-457d-81d4-831ba362a8b4-var-lib\") pod \"ovn-controller-ovs-gh7ls\" (UID: \"fb7ad71b-d500-457d-81d4-831ba362a8b4\") " pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.591441 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb7ad71b-d500-457d-81d4-831ba362a8b4-scripts\") pod \"ovn-controller-ovs-gh7ls\" (UID: \"fb7ad71b-d500-457d-81d4-831ba362a8b4\") " pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.591466 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fb7ad71b-d500-457d-81d4-831ba362a8b4-var-log\") pod \"ovn-controller-ovs-gh7ls\" (UID: \"fb7ad71b-d500-457d-81d4-831ba362a8b4\") " pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.591490 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9dccff31-26fc-43c7-a679-01173d2f34a9-var-run-ovn\") pod \"ovn-controller-tb2c4\" (UID: \"9dccff31-26fc-43c7-a679-01173d2f34a9\") " pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.591516 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrhhs\" (UniqueName: \"kubernetes.io/projected/fb7ad71b-d500-457d-81d4-831ba362a8b4-kube-api-access-xrhhs\") pod \"ovn-controller-ovs-gh7ls\" (UID: \"fb7ad71b-d500-457d-81d4-831ba362a8b4\") " pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.591542 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9v4ft\" (UniqueName: \"kubernetes.io/projected/9dccff31-26fc-43c7-a679-01173d2f34a9-kube-api-access-9v4ft\") pod \"ovn-controller-tb2c4\" (UID: \"9dccff31-26fc-43c7-a679-01173d2f34a9\") " pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.591560 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9dccff31-26fc-43c7-a679-01173d2f34a9-scripts\") pod \"ovn-controller-tb2c4\" (UID: \"9dccff31-26fc-43c7-a679-01173d2f34a9\") " pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.591584 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fb7ad71b-d500-457d-81d4-831ba362a8b4-var-run\") pod \"ovn-controller-ovs-gh7ls\" (UID: \"fb7ad71b-d500-457d-81d4-831ba362a8b4\") " pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.692858 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb7ad71b-d500-457d-81d4-831ba362a8b4-scripts\") pod \"ovn-controller-ovs-gh7ls\" (UID: \"fb7ad71b-d500-457d-81d4-831ba362a8b4\") " pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.692906 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fb7ad71b-d500-457d-81d4-831ba362a8b4-var-log\") pod \"ovn-controller-ovs-gh7ls\" (UID: \"fb7ad71b-d500-457d-81d4-831ba362a8b4\") " pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.692949 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9dccff31-26fc-43c7-a679-01173d2f34a9-var-run-ovn\") pod \"ovn-controller-tb2c4\" (UID: \"9dccff31-26fc-43c7-a679-01173d2f34a9\") " pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.692977 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrhhs\" (UniqueName: \"kubernetes.io/projected/fb7ad71b-d500-457d-81d4-831ba362a8b4-kube-api-access-xrhhs\") pod \"ovn-controller-ovs-gh7ls\" (UID: \"fb7ad71b-d500-457d-81d4-831ba362a8b4\") " pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.693004 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9v4ft\" (UniqueName: \"kubernetes.io/projected/9dccff31-26fc-43c7-a679-01173d2f34a9-kube-api-access-9v4ft\") pod \"ovn-controller-tb2c4\" (UID: \"9dccff31-26fc-43c7-a679-01173d2f34a9\") " pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.693022 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9dccff31-26fc-43c7-a679-01173d2f34a9-scripts\") pod \"ovn-controller-tb2c4\" (UID: \"9dccff31-26fc-43c7-a679-01173d2f34a9\") " pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.693037 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fb7ad71b-d500-457d-81d4-831ba362a8b4-var-run\") pod \"ovn-controller-ovs-gh7ls\" (UID: \"fb7ad71b-d500-457d-81d4-831ba362a8b4\") " pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.693071 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9dccff31-26fc-43c7-a679-01173d2f34a9-var-log-ovn\") pod \"ovn-controller-tb2c4\" (UID: \"9dccff31-26fc-43c7-a679-01173d2f34a9\") " pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.693092 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dccff31-26fc-43c7-a679-01173d2f34a9-ovn-controller-tls-certs\") pod \"ovn-controller-tb2c4\" (UID: \"9dccff31-26fc-43c7-a679-01173d2f34a9\") " pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.693123 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dccff31-26fc-43c7-a679-01173d2f34a9-combined-ca-bundle\") pod \"ovn-controller-tb2c4\" (UID: \"9dccff31-26fc-43c7-a679-01173d2f34a9\") " pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.693143 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fb7ad71b-d500-457d-81d4-831ba362a8b4-etc-ovs\") pod \"ovn-controller-ovs-gh7ls\" (UID: \"fb7ad71b-d500-457d-81d4-831ba362a8b4\") " pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.693170 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9dccff31-26fc-43c7-a679-01173d2f34a9-var-run\") pod \"ovn-controller-tb2c4\" (UID: \"9dccff31-26fc-43c7-a679-01173d2f34a9\") " pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.693183 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fb7ad71b-d500-457d-81d4-831ba362a8b4-var-lib\") pod \"ovn-controller-ovs-gh7ls\" (UID: \"fb7ad71b-d500-457d-81d4-831ba362a8b4\") " pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.693716 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fb7ad71b-d500-457d-81d4-831ba362a8b4-var-lib\") pod \"ovn-controller-ovs-gh7ls\" (UID: \"fb7ad71b-d500-457d-81d4-831ba362a8b4\") " pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.693842 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fb7ad71b-d500-457d-81d4-831ba362a8b4-var-run\") pod \"ovn-controller-ovs-gh7ls\" (UID: \"fb7ad71b-d500-457d-81d4-831ba362a8b4\") " pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.693987 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fb7ad71b-d500-457d-81d4-831ba362a8b4-var-log\") pod \"ovn-controller-ovs-gh7ls\" (UID: \"fb7ad71b-d500-457d-81d4-831ba362a8b4\") " pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.693997 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9dccff31-26fc-43c7-a679-01173d2f34a9-var-log-ovn\") pod \"ovn-controller-tb2c4\" (UID: \"9dccff31-26fc-43c7-a679-01173d2f34a9\") " pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.695767 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9dccff31-26fc-43c7-a679-01173d2f34a9-scripts\") pod \"ovn-controller-tb2c4\" (UID: \"9dccff31-26fc-43c7-a679-01173d2f34a9\") " pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.695879 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9dccff31-26fc-43c7-a679-01173d2f34a9-var-run-ovn\") pod \"ovn-controller-tb2c4\" (UID: \"9dccff31-26fc-43c7-a679-01173d2f34a9\") " pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.696346 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fb7ad71b-d500-457d-81d4-831ba362a8b4-etc-ovs\") pod \"ovn-controller-ovs-gh7ls\" (UID: \"fb7ad71b-d500-457d-81d4-831ba362a8b4\") " pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.696693 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb7ad71b-d500-457d-81d4-831ba362a8b4-scripts\") pod \"ovn-controller-ovs-gh7ls\" (UID: \"fb7ad71b-d500-457d-81d4-831ba362a8b4\") " pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.696767 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9dccff31-26fc-43c7-a679-01173d2f34a9-var-run\") pod \"ovn-controller-tb2c4\" (UID: \"9dccff31-26fc-43c7-a679-01173d2f34a9\") " pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.699474 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dccff31-26fc-43c7-a679-01173d2f34a9-ovn-controller-tls-certs\") pod \"ovn-controller-tb2c4\" (UID: \"9dccff31-26fc-43c7-a679-01173d2f34a9\") " pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.705524 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dccff31-26fc-43c7-a679-01173d2f34a9-combined-ca-bundle\") pod \"ovn-controller-tb2c4\" (UID: \"9dccff31-26fc-43c7-a679-01173d2f34a9\") " pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.709835 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrhhs\" (UniqueName: \"kubernetes.io/projected/fb7ad71b-d500-457d-81d4-831ba362a8b4-kube-api-access-xrhhs\") pod \"ovn-controller-ovs-gh7ls\" (UID: \"fb7ad71b-d500-457d-81d4-831ba362a8b4\") " pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.719613 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9v4ft\" (UniqueName: \"kubernetes.io/projected/9dccff31-26fc-43c7-a679-01173d2f34a9-kube-api-access-9v4ft\") pod \"ovn-controller-tb2c4\" (UID: \"9dccff31-26fc-43c7-a679-01173d2f34a9\") " pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.857379 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:04 crc kubenswrapper[4685]: I1013 09:00:04.867643 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.380365 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.382236 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.386953 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-rktrr" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.387258 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.387475 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.387740 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.388012 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.403312 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.503739 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ffc3833-03fd-40ef-b247-4f1c512c2e8b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.503852 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7ffc3833-03fd-40ef-b247-4f1c512c2e8b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.503897 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ffc3833-03fd-40ef-b247-4f1c512c2e8b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.503960 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.503990 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xl7q\" (UniqueName: \"kubernetes.io/projected/7ffc3833-03fd-40ef-b247-4f1c512c2e8b-kube-api-access-8xl7q\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.504013 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ffc3833-03fd-40ef-b247-4f1c512c2e8b-config\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.504045 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ffc3833-03fd-40ef-b247-4f1c512c2e8b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.504066 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7ffc3833-03fd-40ef-b247-4f1c512c2e8b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.605254 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ffc3833-03fd-40ef-b247-4f1c512c2e8b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.605614 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7ffc3833-03fd-40ef-b247-4f1c512c2e8b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.605665 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ffc3833-03fd-40ef-b247-4f1c512c2e8b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.605789 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7ffc3833-03fd-40ef-b247-4f1c512c2e8b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.606175 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7ffc3833-03fd-40ef-b247-4f1c512c2e8b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.606252 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ffc3833-03fd-40ef-b247-4f1c512c2e8b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.606615 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.606675 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xl7q\" (UniqueName: \"kubernetes.io/projected/7ffc3833-03fd-40ef-b247-4f1c512c2e8b-kube-api-access-8xl7q\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.606705 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ffc3833-03fd-40ef-b247-4f1c512c2e8b-config\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.606767 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7ffc3833-03fd-40ef-b247-4f1c512c2e8b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.606988 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.607460 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ffc3833-03fd-40ef-b247-4f1c512c2e8b-config\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.621265 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ffc3833-03fd-40ef-b247-4f1c512c2e8b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.623058 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ffc3833-03fd-40ef-b247-4f1c512c2e8b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.623326 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ffc3833-03fd-40ef-b247-4f1c512c2e8b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.634260 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xl7q\" (UniqueName: \"kubernetes.io/projected/7ffc3833-03fd-40ef-b247-4f1c512c2e8b-kube-api-access-8xl7q\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.642773 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"7ffc3833-03fd-40ef-b247-4f1c512c2e8b\") " pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.677113 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.678347 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.682181 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.682456 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.682571 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.682505 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-bfdm4" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.699281 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.711687 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.813116 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a3319a1-2f17-43cd-9df1-0697ba10aff8-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.813345 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2a3319a1-2f17-43cd-9df1-0697ba10aff8-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.813430 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a3319a1-2f17-43cd-9df1-0697ba10aff8-config\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.813529 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a3319a1-2f17-43cd-9df1-0697ba10aff8-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.814009 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4g2c\" (UniqueName: \"kubernetes.io/projected/2a3319a1-2f17-43cd-9df1-0697ba10aff8-kube-api-access-f4g2c\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.814081 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2a3319a1-2f17-43cd-9df1-0697ba10aff8-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.814126 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.814153 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a3319a1-2f17-43cd-9df1-0697ba10aff8-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.915103 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4g2c\" (UniqueName: \"kubernetes.io/projected/2a3319a1-2f17-43cd-9df1-0697ba10aff8-kube-api-access-f4g2c\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.915158 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2a3319a1-2f17-43cd-9df1-0697ba10aff8-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.915184 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.915202 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a3319a1-2f17-43cd-9df1-0697ba10aff8-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.915224 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a3319a1-2f17-43cd-9df1-0697ba10aff8-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.915245 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2a3319a1-2f17-43cd-9df1-0697ba10aff8-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.915263 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a3319a1-2f17-43cd-9df1-0697ba10aff8-config\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.915283 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a3319a1-2f17-43cd-9df1-0697ba10aff8-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.915374 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.915810 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2a3319a1-2f17-43cd-9df1-0697ba10aff8-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.916420 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a3319a1-2f17-43cd-9df1-0697ba10aff8-config\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.916755 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2a3319a1-2f17-43cd-9df1-0697ba10aff8-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.919051 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a3319a1-2f17-43cd-9df1-0697ba10aff8-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.925421 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a3319a1-2f17-43cd-9df1-0697ba10aff8-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.928108 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a3319a1-2f17-43cd-9df1-0697ba10aff8-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.931448 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:05 crc kubenswrapper[4685]: I1013 09:00:05.931458 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4g2c\" (UniqueName: \"kubernetes.io/projected/2a3319a1-2f17-43cd-9df1-0697ba10aff8-kube-api-access-f4g2c\") pod \"ovsdbserver-sb-0\" (UID: \"2a3319a1-2f17-43cd-9df1-0697ba10aff8\") " pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:06 crc kubenswrapper[4685]: I1013 09:00:06.023135 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:14 crc kubenswrapper[4685]: E1013 09:00:14.870412 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Oct 13 09:00:14 crc kubenswrapper[4685]: E1013 09:00:14.871113 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v4kzp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(3b0348a6-6814-47f3-919e-e975c01e00e8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 09:00:14 crc kubenswrapper[4685]: E1013 09:00:14.872630 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="3b0348a6-6814-47f3-919e-e975c01e00e8" Oct 13 09:00:15 crc kubenswrapper[4685]: E1013 09:00:15.752844 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 13 09:00:15 crc kubenswrapper[4685]: E1013 09:00:15.753046 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7fgdd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-bztxj_openstack(e7118b17-2973-4bc3-b5b3-027032ff0119): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 09:00:15 crc kubenswrapper[4685]: E1013 09:00:15.754253 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-bztxj" podUID="e7118b17-2973-4bc3-b5b3-027032ff0119" Oct 13 09:00:17 crc kubenswrapper[4685]: E1013 09:00:17.558559 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 13 09:00:17 crc kubenswrapper[4685]: E1013 09:00:17.559134 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sdj26,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-89xzl_openstack(b2134acd-f351-4c06-acc6-b2453a8b0864): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 09:00:17 crc kubenswrapper[4685]: E1013 09:00:17.560350 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-89xzl" podUID="b2134acd-f351-4c06-acc6-b2453a8b0864" Oct 13 09:00:17 crc kubenswrapper[4685]: E1013 09:00:17.587139 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 13 09:00:17 crc kubenswrapper[4685]: E1013 09:00:17.587274 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5hl75,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-znf4k_openstack(9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 09:00:17 crc kubenswrapper[4685]: E1013 09:00:17.588580 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-znf4k" podUID="9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc" Oct 13 09:00:17 crc kubenswrapper[4685]: E1013 09:00:17.605743 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 13 09:00:17 crc kubenswrapper[4685]: E1013 09:00:17.607441 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gx522,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-rcf7p_openstack(a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 09:00:17 crc kubenswrapper[4685]: E1013 09:00:17.608571 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-rcf7p" podUID="a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5" Oct 13 09:00:17 crc kubenswrapper[4685]: I1013 09:00:17.696224 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-bztxj" Oct 13 09:00:17 crc kubenswrapper[4685]: I1013 09:00:17.707463 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fgdd\" (UniqueName: \"kubernetes.io/projected/e7118b17-2973-4bc3-b5b3-027032ff0119-kube-api-access-7fgdd\") pod \"e7118b17-2973-4bc3-b5b3-027032ff0119\" (UID: \"e7118b17-2973-4bc3-b5b3-027032ff0119\") " Oct 13 09:00:17 crc kubenswrapper[4685]: I1013 09:00:17.707587 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7118b17-2973-4bc3-b5b3-027032ff0119-config\") pod \"e7118b17-2973-4bc3-b5b3-027032ff0119\" (UID: \"e7118b17-2973-4bc3-b5b3-027032ff0119\") " Oct 13 09:00:17 crc kubenswrapper[4685]: I1013 09:00:17.708404 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7118b17-2973-4bc3-b5b3-027032ff0119-config" (OuterVolumeSpecName: "config") pod "e7118b17-2973-4bc3-b5b3-027032ff0119" (UID: "e7118b17-2973-4bc3-b5b3-027032ff0119"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:00:17 crc kubenswrapper[4685]: I1013 09:00:17.728114 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7118b17-2973-4bc3-b5b3-027032ff0119-kube-api-access-7fgdd" (OuterVolumeSpecName: "kube-api-access-7fgdd") pod "e7118b17-2973-4bc3-b5b3-027032ff0119" (UID: "e7118b17-2973-4bc3-b5b3-027032ff0119"). InnerVolumeSpecName "kube-api-access-7fgdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:00:17 crc kubenswrapper[4685]: I1013 09:00:17.809745 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fgdd\" (UniqueName: \"kubernetes.io/projected/e7118b17-2973-4bc3-b5b3-027032ff0119-kube-api-access-7fgdd\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:17 crc kubenswrapper[4685]: I1013 09:00:17.809774 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7118b17-2973-4bc3-b5b3-027032ff0119-config\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:18 crc kubenswrapper[4685]: I1013 09:00:18.100191 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339100-h77rw"] Oct 13 09:00:18 crc kubenswrapper[4685]: W1013 09:00:18.111033 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb9b10e0_0158_4c15_9de6_f8f71ecf531b.slice/crio-aba5bbcae489e01ea41a9d026a92a9014669cb501e257e455988cf30e58ce3ee WatchSource:0}: Error finding container aba5bbcae489e01ea41a9d026a92a9014669cb501e257e455988cf30e58ce3ee: Status 404 returned error can't find the container with id aba5bbcae489e01ea41a9d026a92a9014669cb501e257e455988cf30e58ce3ee Oct 13 09:00:18 crc kubenswrapper[4685]: I1013 09:00:18.113564 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 13 09:00:18 crc kubenswrapper[4685]: I1013 09:00:18.322720 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tb2c4"] Oct 13 09:00:18 crc kubenswrapper[4685]: I1013 09:00:18.333673 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 13 09:00:18 crc kubenswrapper[4685]: I1013 09:00:18.381688 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f5482688-fe73-4196-ae4d-a306a1c31545","Type":"ContainerStarted","Data":"affd097dfb44b10c76b141386de4e368da74cb375463d2146f2e005bce26c878"} Oct 13 09:00:18 crc kubenswrapper[4685]: I1013 09:00:18.384174 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"eb9b10e0-0158-4c15-9de6-f8f71ecf531b","Type":"ContainerStarted","Data":"aba5bbcae489e01ea41a9d026a92a9014669cb501e257e455988cf30e58ce3ee"} Oct 13 09:00:18 crc kubenswrapper[4685]: I1013 09:00:18.391093 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339100-h77rw" event={"ID":"edbf1439-6678-4cef-a3ec-ba71bbe1bba6","Type":"ContainerStarted","Data":"2999723a6c8c42cef2091c746f2cd07c61c49b6d337474c5c39902eac0ed587d"} Oct 13 09:00:18 crc kubenswrapper[4685]: I1013 09:00:18.394544 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-bztxj" event={"ID":"e7118b17-2973-4bc3-b5b3-027032ff0119","Type":"ContainerDied","Data":"2c050d6e047cbaa869699ca64fa413d046f73bf8a91ec55f2b5b8b227bd8a035"} Oct 13 09:00:18 crc kubenswrapper[4685]: I1013 09:00:18.394626 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-bztxj" Oct 13 09:00:18 crc kubenswrapper[4685]: W1013 09:00:18.402519 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66b058b9_a4a0_47a9_9a8b_b90e5e9e2b20.slice/crio-35ce1c1bacb38a983136edebbae8954689968686dbfecafe6acaa67da97a8e11 WatchSource:0}: Error finding container 35ce1c1bacb38a983136edebbae8954689968686dbfecafe6acaa67da97a8e11: Status 404 returned error can't find the container with id 35ce1c1bacb38a983136edebbae8954689968686dbfecafe6acaa67da97a8e11 Oct 13 09:00:18 crc kubenswrapper[4685]: E1013 09:00:18.402526 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-rcf7p" podUID="a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5" Oct 13 09:00:18 crc kubenswrapper[4685]: E1013 09:00:18.413994 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-znf4k" podUID="9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc" Oct 13 09:00:18 crc kubenswrapper[4685]: I1013 09:00:18.547389 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 13 09:00:18 crc kubenswrapper[4685]: I1013 09:00:18.562597 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-bztxj"] Oct 13 09:00:18 crc kubenswrapper[4685]: I1013 09:00:18.575382 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-bztxj"] Oct 13 09:00:18 crc kubenswrapper[4685]: I1013 09:00:18.867885 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-89xzl" Oct 13 09:00:19 crc kubenswrapper[4685]: I1013 09:00:19.038030 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2134acd-f351-4c06-acc6-b2453a8b0864-dns-svc\") pod \"b2134acd-f351-4c06-acc6-b2453a8b0864\" (UID: \"b2134acd-f351-4c06-acc6-b2453a8b0864\") " Oct 13 09:00:19 crc kubenswrapper[4685]: I1013 09:00:19.038102 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdj26\" (UniqueName: \"kubernetes.io/projected/b2134acd-f351-4c06-acc6-b2453a8b0864-kube-api-access-sdj26\") pod \"b2134acd-f351-4c06-acc6-b2453a8b0864\" (UID: \"b2134acd-f351-4c06-acc6-b2453a8b0864\") " Oct 13 09:00:19 crc kubenswrapper[4685]: I1013 09:00:19.038244 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2134acd-f351-4c06-acc6-b2453a8b0864-config\") pod \"b2134acd-f351-4c06-acc6-b2453a8b0864\" (UID: \"b2134acd-f351-4c06-acc6-b2453a8b0864\") " Oct 13 09:00:19 crc kubenswrapper[4685]: I1013 09:00:19.038714 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2134acd-f351-4c06-acc6-b2453a8b0864-config" (OuterVolumeSpecName: "config") pod "b2134acd-f351-4c06-acc6-b2453a8b0864" (UID: "b2134acd-f351-4c06-acc6-b2453a8b0864"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:00:19 crc kubenswrapper[4685]: I1013 09:00:19.039346 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2134acd-f351-4c06-acc6-b2453a8b0864-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b2134acd-f351-4c06-acc6-b2453a8b0864" (UID: "b2134acd-f351-4c06-acc6-b2453a8b0864"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:00:19 crc kubenswrapper[4685]: I1013 09:00:19.043152 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2134acd-f351-4c06-acc6-b2453a8b0864-kube-api-access-sdj26" (OuterVolumeSpecName: "kube-api-access-sdj26") pod "b2134acd-f351-4c06-acc6-b2453a8b0864" (UID: "b2134acd-f351-4c06-acc6-b2453a8b0864"). InnerVolumeSpecName "kube-api-access-sdj26". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:00:19 crc kubenswrapper[4685]: I1013 09:00:19.139576 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2134acd-f351-4c06-acc6-b2453a8b0864-config\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:19 crc kubenswrapper[4685]: I1013 09:00:19.139609 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2134acd-f351-4c06-acc6-b2453a8b0864-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:19 crc kubenswrapper[4685]: I1013 09:00:19.139618 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdj26\" (UniqueName: \"kubernetes.io/projected/b2134acd-f351-4c06-acc6-b2453a8b0864-kube-api-access-sdj26\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:19 crc kubenswrapper[4685]: I1013 09:00:19.406860 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"fb5d643a-96c6-4fca-904d-3b4a71e9630c","Type":"ContainerStarted","Data":"e4defc1e5f088109f7e5b8500f3418fdaa55e866f01f7b91276116d3525c47ee"} Oct 13 09:00:19 crc kubenswrapper[4685]: I1013 09:00:19.410614 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-89xzl" event={"ID":"b2134acd-f351-4c06-acc6-b2453a8b0864","Type":"ContainerDied","Data":"ca6f5cfa32ec99b7c7675bdc33362168a626c01ef78603de30041fa488a8cda6"} Oct 13 09:00:19 crc kubenswrapper[4685]: I1013 09:00:19.410651 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-89xzl" Oct 13 09:00:19 crc kubenswrapper[4685]: I1013 09:00:19.413158 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20","Type":"ContainerStarted","Data":"1d77c015c2997227a86e7faa0f74eb5c8e31c2ac0703a5685575da6a8757f89f"} Oct 13 09:00:19 crc kubenswrapper[4685]: I1013 09:00:19.413239 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20","Type":"ContainerStarted","Data":"35ce1c1bacb38a983136edebbae8954689968686dbfecafe6acaa67da97a8e11"} Oct 13 09:00:19 crc kubenswrapper[4685]: I1013 09:00:19.415413 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tb2c4" event={"ID":"9dccff31-26fc-43c7-a679-01173d2f34a9","Type":"ContainerStarted","Data":"5628929e3db98508af10de42943a1871a7b14569e79be741e23851aa1ac0cd8c"} Oct 13 09:00:19 crc kubenswrapper[4685]: I1013 09:00:19.421355 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3b0348a6-6814-47f3-919e-e975c01e00e8","Type":"ContainerStarted","Data":"a8c575675b8a8240f9a9cce662be0dbf91b784093b2ec63f6e3d873f5d79efc2"} Oct 13 09:00:19 crc kubenswrapper[4685]: I1013 09:00:19.424959 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b7f19239-a2d9-49b5-80b0-43543daf1787","Type":"ContainerStarted","Data":"a6d714e0fc56e7256e93150fb3ec702f4835022733af8f3f1d12eecd1ec2d20b"} Oct 13 09:00:19 crc kubenswrapper[4685]: I1013 09:00:19.427245 4685 generic.go:334] "Generic (PLEG): container finished" podID="edbf1439-6678-4cef-a3ec-ba71bbe1bba6" containerID="0afedaf401d57b82882990da486231433d3b6bb69c611ef47e1b0e99b494766f" exitCode=0 Oct 13 09:00:19 crc kubenswrapper[4685]: I1013 09:00:19.427526 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339100-h77rw" event={"ID":"edbf1439-6678-4cef-a3ec-ba71bbe1bba6","Type":"ContainerDied","Data":"0afedaf401d57b82882990da486231433d3b6bb69c611ef47e1b0e99b494766f"} Oct 13 09:00:19 crc kubenswrapper[4685]: I1013 09:00:19.517754 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7118b17-2973-4bc3-b5b3-027032ff0119" path="/var/lib/kubelet/pods/e7118b17-2973-4bc3-b5b3-027032ff0119/volumes" Oct 13 09:00:19 crc kubenswrapper[4685]: I1013 09:00:19.518420 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-89xzl"] Oct 13 09:00:19 crc kubenswrapper[4685]: I1013 09:00:19.534097 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-89xzl"] Oct 13 09:00:20 crc kubenswrapper[4685]: I1013 09:00:20.239941 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-gh7ls"] Oct 13 09:00:20 crc kubenswrapper[4685]: I1013 09:00:20.789060 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339100-h77rw" Oct 13 09:00:20 crc kubenswrapper[4685]: I1013 09:00:20.866389 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/edbf1439-6678-4cef-a3ec-ba71bbe1bba6-secret-volume\") pod \"edbf1439-6678-4cef-a3ec-ba71bbe1bba6\" (UID: \"edbf1439-6678-4cef-a3ec-ba71bbe1bba6\") " Oct 13 09:00:20 crc kubenswrapper[4685]: I1013 09:00:20.866457 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/edbf1439-6678-4cef-a3ec-ba71bbe1bba6-config-volume\") pod \"edbf1439-6678-4cef-a3ec-ba71bbe1bba6\" (UID: \"edbf1439-6678-4cef-a3ec-ba71bbe1bba6\") " Oct 13 09:00:20 crc kubenswrapper[4685]: I1013 09:00:20.866568 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n68ss\" (UniqueName: \"kubernetes.io/projected/edbf1439-6678-4cef-a3ec-ba71bbe1bba6-kube-api-access-n68ss\") pod \"edbf1439-6678-4cef-a3ec-ba71bbe1bba6\" (UID: \"edbf1439-6678-4cef-a3ec-ba71bbe1bba6\") " Oct 13 09:00:20 crc kubenswrapper[4685]: I1013 09:00:20.868325 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edbf1439-6678-4cef-a3ec-ba71bbe1bba6-config-volume" (OuterVolumeSpecName: "config-volume") pod "edbf1439-6678-4cef-a3ec-ba71bbe1bba6" (UID: "edbf1439-6678-4cef-a3ec-ba71bbe1bba6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:00:20 crc kubenswrapper[4685]: I1013 09:00:20.884662 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edbf1439-6678-4cef-a3ec-ba71bbe1bba6-kube-api-access-n68ss" (OuterVolumeSpecName: "kube-api-access-n68ss") pod "edbf1439-6678-4cef-a3ec-ba71bbe1bba6" (UID: "edbf1439-6678-4cef-a3ec-ba71bbe1bba6"). InnerVolumeSpecName "kube-api-access-n68ss". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:00:20 crc kubenswrapper[4685]: I1013 09:00:20.909159 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edbf1439-6678-4cef-a3ec-ba71bbe1bba6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "edbf1439-6678-4cef-a3ec-ba71bbe1bba6" (UID: "edbf1439-6678-4cef-a3ec-ba71bbe1bba6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:00:20 crc kubenswrapper[4685]: I1013 09:00:20.921173 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 13 09:00:20 crc kubenswrapper[4685]: W1013 09:00:20.930353 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a3319a1_2f17_43cd_9df1_0697ba10aff8.slice/crio-ac2f31ad895cd9c163ca60e7e94781b00bc0c690176cb6b4f0dce09d7e20c81b WatchSource:0}: Error finding container ac2f31ad895cd9c163ca60e7e94781b00bc0c690176cb6b4f0dce09d7e20c81b: Status 404 returned error can't find the container with id ac2f31ad895cd9c163ca60e7e94781b00bc0c690176cb6b4f0dce09d7e20c81b Oct 13 09:00:20 crc kubenswrapper[4685]: I1013 09:00:20.968856 4685 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/edbf1439-6678-4cef-a3ec-ba71bbe1bba6-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:20 crc kubenswrapper[4685]: I1013 09:00:20.968889 4685 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/edbf1439-6678-4cef-a3ec-ba71bbe1bba6-config-volume\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:20 crc kubenswrapper[4685]: I1013 09:00:20.968901 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n68ss\" (UniqueName: \"kubernetes.io/projected/edbf1439-6678-4cef-a3ec-ba71bbe1bba6-kube-api-access-n68ss\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:21 crc kubenswrapper[4685]: I1013 09:00:21.102282 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 13 09:00:21 crc kubenswrapper[4685]: W1013 09:00:21.129926 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ffc3833_03fd_40ef_b247_4f1c512c2e8b.slice/crio-8c25909d8b4e0462674309ae681e4a1bc5fccd31ff42201a9fa49fc17bf198e1 WatchSource:0}: Error finding container 8c25909d8b4e0462674309ae681e4a1bc5fccd31ff42201a9fa49fc17bf198e1: Status 404 returned error can't find the container with id 8c25909d8b4e0462674309ae681e4a1bc5fccd31ff42201a9fa49fc17bf198e1 Oct 13 09:00:21 crc kubenswrapper[4685]: I1013 09:00:21.444976 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gh7ls" event={"ID":"fb7ad71b-d500-457d-81d4-831ba362a8b4","Type":"ContainerStarted","Data":"de9bf85aaab70842453330c309182d0b377545980c18f5a0a192a4e26f8104e1"} Oct 13 09:00:21 crc kubenswrapper[4685]: I1013 09:00:21.447531 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2a3319a1-2f17-43cd-9df1-0697ba10aff8","Type":"ContainerStarted","Data":"ac2f31ad895cd9c163ca60e7e94781b00bc0c690176cb6b4f0dce09d7e20c81b"} Oct 13 09:00:21 crc kubenswrapper[4685]: I1013 09:00:21.449705 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339100-h77rw" event={"ID":"edbf1439-6678-4cef-a3ec-ba71bbe1bba6","Type":"ContainerDied","Data":"2999723a6c8c42cef2091c746f2cd07c61c49b6d337474c5c39902eac0ed587d"} Oct 13 09:00:21 crc kubenswrapper[4685]: I1013 09:00:21.449753 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2999723a6c8c42cef2091c746f2cd07c61c49b6d337474c5c39902eac0ed587d" Oct 13 09:00:21 crc kubenswrapper[4685]: I1013 09:00:21.449720 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339100-h77rw" Oct 13 09:00:21 crc kubenswrapper[4685]: I1013 09:00:21.451483 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"7ffc3833-03fd-40ef-b247-4f1c512c2e8b","Type":"ContainerStarted","Data":"8c25909d8b4e0462674309ae681e4a1bc5fccd31ff42201a9fa49fc17bf198e1"} Oct 13 09:00:21 crc kubenswrapper[4685]: I1013 09:00:21.514667 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2134acd-f351-4c06-acc6-b2453a8b0864" path="/var/lib/kubelet/pods/b2134acd-f351-4c06-acc6-b2453a8b0864/volumes" Oct 13 09:00:24 crc kubenswrapper[4685]: I1013 09:00:24.482056 4685 generic.go:334] "Generic (PLEG): container finished" podID="66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20" containerID="1d77c015c2997227a86e7faa0f74eb5c8e31c2ac0703a5685575da6a8757f89f" exitCode=0 Oct 13 09:00:24 crc kubenswrapper[4685]: I1013 09:00:24.482189 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20","Type":"ContainerDied","Data":"1d77c015c2997227a86e7faa0f74eb5c8e31c2ac0703a5685575da6a8757f89f"} Oct 13 09:00:24 crc kubenswrapper[4685]: I1013 09:00:24.493396 4685 generic.go:334] "Generic (PLEG): container finished" podID="f5482688-fe73-4196-ae4d-a306a1c31545" containerID="affd097dfb44b10c76b141386de4e368da74cb375463d2146f2e005bce26c878" exitCode=0 Oct 13 09:00:24 crc kubenswrapper[4685]: I1013 09:00:24.493428 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f5482688-fe73-4196-ae4d-a306a1c31545","Type":"ContainerDied","Data":"affd097dfb44b10c76b141386de4e368da74cb375463d2146f2e005bce26c878"} Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.440156 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-t646z"] Oct 13 09:00:25 crc kubenswrapper[4685]: E1013 09:00:25.440804 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edbf1439-6678-4cef-a3ec-ba71bbe1bba6" containerName="collect-profiles" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.440816 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="edbf1439-6678-4cef-a3ec-ba71bbe1bba6" containerName="collect-profiles" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.441020 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="edbf1439-6678-4cef-a3ec-ba71bbe1bba6" containerName="collect-profiles" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.441569 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-t646z" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.449278 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.452938 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-t646z"] Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.551770 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/85cf8998-48c3-49b9-9a68-d5765b785df4-ovn-rundir\") pod \"ovn-controller-metrics-t646z\" (UID: \"85cf8998-48c3-49b9-9a68-d5765b785df4\") " pod="openstack/ovn-controller-metrics-t646z" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.551827 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hxz5\" (UniqueName: \"kubernetes.io/projected/85cf8998-48c3-49b9-9a68-d5765b785df4-kube-api-access-2hxz5\") pod \"ovn-controller-metrics-t646z\" (UID: \"85cf8998-48c3-49b9-9a68-d5765b785df4\") " pod="openstack/ovn-controller-metrics-t646z" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.551886 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85cf8998-48c3-49b9-9a68-d5765b785df4-config\") pod \"ovn-controller-metrics-t646z\" (UID: \"85cf8998-48c3-49b9-9a68-d5765b785df4\") " pod="openstack/ovn-controller-metrics-t646z" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.551936 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/85cf8998-48c3-49b9-9a68-d5765b785df4-ovs-rundir\") pod \"ovn-controller-metrics-t646z\" (UID: \"85cf8998-48c3-49b9-9a68-d5765b785df4\") " pod="openstack/ovn-controller-metrics-t646z" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.551996 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85cf8998-48c3-49b9-9a68-d5765b785df4-combined-ca-bundle\") pod \"ovn-controller-metrics-t646z\" (UID: \"85cf8998-48c3-49b9-9a68-d5765b785df4\") " pod="openstack/ovn-controller-metrics-t646z" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.552024 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/85cf8998-48c3-49b9-9a68-d5765b785df4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-t646z\" (UID: \"85cf8998-48c3-49b9-9a68-d5765b785df4\") " pod="openstack/ovn-controller-metrics-t646z" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.660807 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/85cf8998-48c3-49b9-9a68-d5765b785df4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-t646z\" (UID: \"85cf8998-48c3-49b9-9a68-d5765b785df4\") " pod="openstack/ovn-controller-metrics-t646z" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.660860 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/85cf8998-48c3-49b9-9a68-d5765b785df4-ovn-rundir\") pod \"ovn-controller-metrics-t646z\" (UID: \"85cf8998-48c3-49b9-9a68-d5765b785df4\") " pod="openstack/ovn-controller-metrics-t646z" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.660890 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hxz5\" (UniqueName: \"kubernetes.io/projected/85cf8998-48c3-49b9-9a68-d5765b785df4-kube-api-access-2hxz5\") pod \"ovn-controller-metrics-t646z\" (UID: \"85cf8998-48c3-49b9-9a68-d5765b785df4\") " pod="openstack/ovn-controller-metrics-t646z" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.660949 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85cf8998-48c3-49b9-9a68-d5765b785df4-config\") pod \"ovn-controller-metrics-t646z\" (UID: \"85cf8998-48c3-49b9-9a68-d5765b785df4\") " pod="openstack/ovn-controller-metrics-t646z" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.660974 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/85cf8998-48c3-49b9-9a68-d5765b785df4-ovs-rundir\") pod \"ovn-controller-metrics-t646z\" (UID: \"85cf8998-48c3-49b9-9a68-d5765b785df4\") " pod="openstack/ovn-controller-metrics-t646z" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.661018 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85cf8998-48c3-49b9-9a68-d5765b785df4-combined-ca-bundle\") pod \"ovn-controller-metrics-t646z\" (UID: \"85cf8998-48c3-49b9-9a68-d5765b785df4\") " pod="openstack/ovn-controller-metrics-t646z" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.664391 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/85cf8998-48c3-49b9-9a68-d5765b785df4-ovn-rundir\") pod \"ovn-controller-metrics-t646z\" (UID: \"85cf8998-48c3-49b9-9a68-d5765b785df4\") " pod="openstack/ovn-controller-metrics-t646z" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.665196 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85cf8998-48c3-49b9-9a68-d5765b785df4-config\") pod \"ovn-controller-metrics-t646z\" (UID: \"85cf8998-48c3-49b9-9a68-d5765b785df4\") " pod="openstack/ovn-controller-metrics-t646z" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.665270 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/85cf8998-48c3-49b9-9a68-d5765b785df4-ovs-rundir\") pod \"ovn-controller-metrics-t646z\" (UID: \"85cf8998-48c3-49b9-9a68-d5765b785df4\") " pod="openstack/ovn-controller-metrics-t646z" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.676694 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85cf8998-48c3-49b9-9a68-d5765b785df4-combined-ca-bundle\") pod \"ovn-controller-metrics-t646z\" (UID: \"85cf8998-48c3-49b9-9a68-d5765b785df4\") " pod="openstack/ovn-controller-metrics-t646z" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.678352 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/85cf8998-48c3-49b9-9a68-d5765b785df4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-t646z\" (UID: \"85cf8998-48c3-49b9-9a68-d5765b785df4\") " pod="openstack/ovn-controller-metrics-t646z" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.695331 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-znf4k"] Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.714056 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hxz5\" (UniqueName: \"kubernetes.io/projected/85cf8998-48c3-49b9-9a68-d5765b785df4-kube-api-access-2hxz5\") pod \"ovn-controller-metrics-t646z\" (UID: \"85cf8998-48c3-49b9-9a68-d5765b785df4\") " pod="openstack/ovn-controller-metrics-t646z" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.728347 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-szmsc"] Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.729668 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.733674 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.773672 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-t646z" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.792139 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-szmsc"] Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.866766 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q56cd\" (UniqueName: \"kubernetes.io/projected/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-kube-api-access-q56cd\") pod \"dnsmasq-dns-5bf47b49b7-szmsc\" (UID: \"9ce22efc-b48d-4d42-a01a-31f71c53c3ab\") " pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.867188 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-szmsc\" (UID: \"9ce22efc-b48d-4d42-a01a-31f71c53c3ab\") " pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.867232 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-szmsc\" (UID: \"9ce22efc-b48d-4d42-a01a-31f71c53c3ab\") " pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.867260 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-config\") pod \"dnsmasq-dns-5bf47b49b7-szmsc\" (UID: \"9ce22efc-b48d-4d42-a01a-31f71c53c3ab\") " pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.945270 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-rcf7p"] Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.973837 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-config\") pod \"dnsmasq-dns-5bf47b49b7-szmsc\" (UID: \"9ce22efc-b48d-4d42-a01a-31f71c53c3ab\") " pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.973981 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q56cd\" (UniqueName: \"kubernetes.io/projected/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-kube-api-access-q56cd\") pod \"dnsmasq-dns-5bf47b49b7-szmsc\" (UID: \"9ce22efc-b48d-4d42-a01a-31f71c53c3ab\") " pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.974021 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-szmsc\" (UID: \"9ce22efc-b48d-4d42-a01a-31f71c53c3ab\") " pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.974038 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-szmsc\" (UID: \"9ce22efc-b48d-4d42-a01a-31f71c53c3ab\") " pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.974765 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-szmsc\" (UID: \"9ce22efc-b48d-4d42-a01a-31f71c53c3ab\") " pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.975272 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-config\") pod \"dnsmasq-dns-5bf47b49b7-szmsc\" (UID: \"9ce22efc-b48d-4d42-a01a-31f71c53c3ab\") " pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.975754 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-szmsc\" (UID: \"9ce22efc-b48d-4d42-a01a-31f71c53c3ab\") " pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.977854 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-dl7rb"] Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.982935 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-dl7rb" Oct 13 09:00:25 crc kubenswrapper[4685]: I1013 09:00:25.989611 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.008070 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q56cd\" (UniqueName: \"kubernetes.io/projected/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-kube-api-access-q56cd\") pod \"dnsmasq-dns-5bf47b49b7-szmsc\" (UID: \"9ce22efc-b48d-4d42-a01a-31f71c53c3ab\") " pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.033441 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-dl7rb"] Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.075285 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-config\") pod \"dnsmasq-dns-8554648995-dl7rb\" (UID: \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\") " pod="openstack/dnsmasq-dns-8554648995-dl7rb" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.075671 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-dl7rb\" (UID: \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\") " pod="openstack/dnsmasq-dns-8554648995-dl7rb" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.075693 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x25b2\" (UniqueName: \"kubernetes.io/projected/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-kube-api-access-x25b2\") pod \"dnsmasq-dns-8554648995-dl7rb\" (UID: \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\") " pod="openstack/dnsmasq-dns-8554648995-dl7rb" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.075729 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-dns-svc\") pod \"dnsmasq-dns-8554648995-dl7rb\" (UID: \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\") " pod="openstack/dnsmasq-dns-8554648995-dl7rb" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.075744 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-dl7rb\" (UID: \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\") " pod="openstack/dnsmasq-dns-8554648995-dl7rb" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.081129 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.177648 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-dns-svc\") pod \"dnsmasq-dns-8554648995-dl7rb\" (UID: \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\") " pod="openstack/dnsmasq-dns-8554648995-dl7rb" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.177685 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-dl7rb\" (UID: \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\") " pod="openstack/dnsmasq-dns-8554648995-dl7rb" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.177716 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-config\") pod \"dnsmasq-dns-8554648995-dl7rb\" (UID: \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\") " pod="openstack/dnsmasq-dns-8554648995-dl7rb" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.177810 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-dl7rb\" (UID: \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\") " pod="openstack/dnsmasq-dns-8554648995-dl7rb" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.177830 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x25b2\" (UniqueName: \"kubernetes.io/projected/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-kube-api-access-x25b2\") pod \"dnsmasq-dns-8554648995-dl7rb\" (UID: \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\") " pod="openstack/dnsmasq-dns-8554648995-dl7rb" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.179230 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-config\") pod \"dnsmasq-dns-8554648995-dl7rb\" (UID: \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\") " pod="openstack/dnsmasq-dns-8554648995-dl7rb" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.179342 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-dl7rb\" (UID: \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\") " pod="openstack/dnsmasq-dns-8554648995-dl7rb" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.179600 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-dl7rb\" (UID: \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\") " pod="openstack/dnsmasq-dns-8554648995-dl7rb" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.180376 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-dns-svc\") pod \"dnsmasq-dns-8554648995-dl7rb\" (UID: \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\") " pod="openstack/dnsmasq-dns-8554648995-dl7rb" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.202041 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x25b2\" (UniqueName: \"kubernetes.io/projected/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-kube-api-access-x25b2\") pod \"dnsmasq-dns-8554648995-dl7rb\" (UID: \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\") " pod="openstack/dnsmasq-dns-8554648995-dl7rb" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.488761 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-t646z"] Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.519403 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"eb9b10e0-0158-4c15-9de6-f8f71ecf531b","Type":"ContainerStarted","Data":"26a6ac5b5aa96d798c66b5a6042ee923909475d47cacf0f29744e11debdcd956"} Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.520078 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.522623 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-znf4k" event={"ID":"9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc","Type":"ContainerDied","Data":"8609e32c6cf331898dc5abb70a4fc99f765a56eab90d257388f9c51139de47bc"} Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.522647 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8609e32c6cf331898dc5abb70a4fc99f765a56eab90d257388f9c51139de47bc" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.524539 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"fb5d643a-96c6-4fca-904d-3b4a71e9630c","Type":"ContainerStarted","Data":"707c3db51ff6b756b8c9b9468695e51823f28bb891021161af196a5db665de4c"} Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.525215 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.531229 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20","Type":"ContainerStarted","Data":"da4278b12727532cd2023318d86bba0d72e4c4194849580ca148a7d6f338b184"} Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.538434 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=20.05270282 podStartE2EDuration="27.538415756s" podCreationTimestamp="2025-10-13 08:59:59 +0000 UTC" firstStartedPulling="2025-10-13 09:00:18.122056555 +0000 UTC m=+943.269932316" lastFinishedPulling="2025-10-13 09:00:25.607769491 +0000 UTC m=+950.755645252" observedRunningTime="2025-10-13 09:00:26.537018827 +0000 UTC m=+951.684894598" watchObservedRunningTime="2025-10-13 09:00:26.538415756 +0000 UTC m=+951.686291517" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.554276 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f5482688-fe73-4196-ae4d-a306a1c31545","Type":"ContainerStarted","Data":"25384cab37ea723cbecc00ae949418d4955eddd45b44ef7e570cf8cbc08f74d9"} Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.556291 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=22.565173059 podStartE2EDuration="29.556279835s" podCreationTimestamp="2025-10-13 08:59:57 +0000 UTC" firstStartedPulling="2025-10-13 09:00:18.56614043 +0000 UTC m=+943.714016191" lastFinishedPulling="2025-10-13 09:00:25.557247206 +0000 UTC m=+950.705122967" observedRunningTime="2025-10-13 09:00:26.555133054 +0000 UTC m=+951.703008815" watchObservedRunningTime="2025-10-13 09:00:26.556279835 +0000 UTC m=+951.704155596" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.572949 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-rcf7p" event={"ID":"a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5","Type":"ContainerDied","Data":"804c70c18f0b8a9e049ccf3b0e9618018a149c9e039412d35dc0ae91a2fa1a4d"} Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.572991 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="804c70c18f0b8a9e049ccf3b0e9618018a149c9e039412d35dc0ae91a2fa1a4d" Oct 13 09:00:26 crc kubenswrapper[4685]: W1013 09:00:26.579718 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85cf8998_48c3_49b9_9a68_d5765b785df4.slice/crio-5a6e078511642336b2df1cd92dc4412b363ce587f82344d7d7d4e05a4383dddd WatchSource:0}: Error finding container 5a6e078511642336b2df1cd92dc4412b363ce587f82344d7d7d4e05a4383dddd: Status 404 returned error can't find the container with id 5a6e078511642336b2df1cd92dc4412b363ce587f82344d7d7d4e05a4383dddd Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.591306 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=30.591288303 podStartE2EDuration="30.591288303s" podCreationTimestamp="2025-10-13 08:59:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:00:26.589154076 +0000 UTC m=+951.737029837" watchObservedRunningTime="2025-10-13 09:00:26.591288303 +0000 UTC m=+951.739164064" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.620898 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=17.286624079 podStartE2EDuration="31.620878884s" podCreationTimestamp="2025-10-13 08:59:55 +0000 UTC" firstStartedPulling="2025-10-13 09:00:03.327903141 +0000 UTC m=+928.475778902" lastFinishedPulling="2025-10-13 09:00:17.662157946 +0000 UTC m=+942.810033707" observedRunningTime="2025-10-13 09:00:26.617640776 +0000 UTC m=+951.765516547" watchObservedRunningTime="2025-10-13 09:00:26.620878884 +0000 UTC m=+951.768754645" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.637429 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-dl7rb" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.646090 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-znf4k" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.650032 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-szmsc"] Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.723078 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-rcf7p" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.792592 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5-dns-svc\") pod \"a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5\" (UID: \"a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5\") " Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.792963 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc-dns-svc\") pod \"9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc\" (UID: \"9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc\") " Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.793034 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc-config\") pod \"9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc\" (UID: \"9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc\") " Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.793060 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hl75\" (UniqueName: \"kubernetes.io/projected/9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc-kube-api-access-5hl75\") pod \"9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc\" (UID: \"9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc\") " Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.793118 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5-config\") pod \"a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5\" (UID: \"a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5\") " Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.793221 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gx522\" (UniqueName: \"kubernetes.io/projected/a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5-kube-api-access-gx522\") pod \"a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5\" (UID: \"a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5\") " Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.796867 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5-config" (OuterVolumeSpecName: "config") pod "a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5" (UID: "a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.800425 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5" (UID: "a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.801409 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc-config" (OuterVolumeSpecName: "config") pod "9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc" (UID: "9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.802769 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc" (UID: "9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.831101 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc-kube-api-access-5hl75" (OuterVolumeSpecName: "kube-api-access-5hl75") pod "9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc" (UID: "9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc"). InnerVolumeSpecName "kube-api-access-5hl75". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.831663 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5-kube-api-access-gx522" (OuterVolumeSpecName: "kube-api-access-gx522") pod "a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5" (UID: "a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5"). InnerVolumeSpecName "kube-api-access-gx522". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.895499 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gx522\" (UniqueName: \"kubernetes.io/projected/a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5-kube-api-access-gx522\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.895529 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.895539 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.895546 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc-config\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.895555 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hl75\" (UniqueName: \"kubernetes.io/projected/9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc-kube-api-access-5hl75\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:26 crc kubenswrapper[4685]: I1013 09:00:26.895563 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5-config\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:27 crc kubenswrapper[4685]: I1013 09:00:27.247807 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-dl7rb"] Oct 13 09:00:27 crc kubenswrapper[4685]: I1013 09:00:27.583050 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tb2c4" event={"ID":"9dccff31-26fc-43c7-a679-01173d2f34a9","Type":"ContainerStarted","Data":"54763bd15f9f6ecc33884114178b9a0451e34710ccbe0cfc57d2c1984bd17c01"} Oct 13 09:00:27 crc kubenswrapper[4685]: I1013 09:00:27.583373 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-tb2c4" Oct 13 09:00:27 crc kubenswrapper[4685]: I1013 09:00:27.586408 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" event={"ID":"9ce22efc-b48d-4d42-a01a-31f71c53c3ab","Type":"ContainerStarted","Data":"63fc65662155bc9c74d683ff726f68ca952d2e0c78e81c7b946bcc50dba4da94"} Oct 13 09:00:27 crc kubenswrapper[4685]: I1013 09:00:27.586430 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" event={"ID":"9ce22efc-b48d-4d42-a01a-31f71c53c3ab","Type":"ContainerStarted","Data":"0357011fddf91c251e92980dbcfac2fa4f7c87faa594714b40ad5587e7e75edc"} Oct 13 09:00:27 crc kubenswrapper[4685]: I1013 09:00:27.590316 4685 generic.go:334] "Generic (PLEG): container finished" podID="fb7ad71b-d500-457d-81d4-831ba362a8b4" containerID="c685e6a184ff94ae883804c7b30319dcae9a179832ce176782e78792da447191" exitCode=0 Oct 13 09:00:27 crc kubenswrapper[4685]: I1013 09:00:27.590367 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gh7ls" event={"ID":"fb7ad71b-d500-457d-81d4-831ba362a8b4","Type":"ContainerDied","Data":"c685e6a184ff94ae883804c7b30319dcae9a179832ce176782e78792da447191"} Oct 13 09:00:27 crc kubenswrapper[4685]: I1013 09:00:27.594711 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-dl7rb" event={"ID":"86735b6c-f0ad-47b7-9af5-f07264a2fbcd","Type":"ContainerStarted","Data":"3a7920f25591dd079e23c2fd29360fb24d151297f61d242c53353415e0ea47f0"} Oct 13 09:00:27 crc kubenswrapper[4685]: I1013 09:00:27.600811 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2a3319a1-2f17-43cd-9df1-0697ba10aff8","Type":"ContainerStarted","Data":"2135794403bea9e3293c3d3e2e416edc8b41abf5eeb3e3c0137b74a61079fd30"} Oct 13 09:00:27 crc kubenswrapper[4685]: I1013 09:00:27.620006 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-t646z" event={"ID":"85cf8998-48c3-49b9-9a68-d5765b785df4","Type":"ContainerStarted","Data":"5a6e078511642336b2df1cd92dc4412b363ce587f82344d7d7d4e05a4383dddd"} Oct 13 09:00:27 crc kubenswrapper[4685]: I1013 09:00:27.631082 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-znf4k" Oct 13 09:00:27 crc kubenswrapper[4685]: I1013 09:00:27.631491 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-tb2c4" podStartSLOduration=16.303258627 podStartE2EDuration="23.631477089s" podCreationTimestamp="2025-10-13 09:00:04 +0000 UTC" firstStartedPulling="2025-10-13 09:00:18.416154181 +0000 UTC m=+943.564029942" lastFinishedPulling="2025-10-13 09:00:25.744372643 +0000 UTC m=+950.892248404" observedRunningTime="2025-10-13 09:00:27.607846021 +0000 UTC m=+952.755721802" watchObservedRunningTime="2025-10-13 09:00:27.631477089 +0000 UTC m=+952.779352850" Oct 13 09:00:27 crc kubenswrapper[4685]: I1013 09:00:27.632321 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-rcf7p" Oct 13 09:00:27 crc kubenswrapper[4685]: I1013 09:00:27.632426 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"7ffc3833-03fd-40ef-b247-4f1c512c2e8b","Type":"ContainerStarted","Data":"be802f83db376f529f93f6b8283339d0722f7c03736d5146b6eda95588ac6178"} Oct 13 09:00:27 crc kubenswrapper[4685]: I1013 09:00:27.693173 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 13 09:00:27 crc kubenswrapper[4685]: I1013 09:00:27.693900 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 13 09:00:27 crc kubenswrapper[4685]: I1013 09:00:27.714396 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-rcf7p"] Oct 13 09:00:27 crc kubenswrapper[4685]: I1013 09:00:27.746897 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-rcf7p"] Oct 13 09:00:27 crc kubenswrapper[4685]: I1013 09:00:27.764051 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-znf4k"] Oct 13 09:00:27 crc kubenswrapper[4685]: I1013 09:00:27.772925 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-znf4k"] Oct 13 09:00:28 crc kubenswrapper[4685]: I1013 09:00:28.642223 4685 generic.go:334] "Generic (PLEG): container finished" podID="9ce22efc-b48d-4d42-a01a-31f71c53c3ab" containerID="63fc65662155bc9c74d683ff726f68ca952d2e0c78e81c7b946bcc50dba4da94" exitCode=0 Oct 13 09:00:28 crc kubenswrapper[4685]: I1013 09:00:28.642314 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" event={"ID":"9ce22efc-b48d-4d42-a01a-31f71c53c3ab","Type":"ContainerDied","Data":"63fc65662155bc9c74d683ff726f68ca952d2e0c78e81c7b946bcc50dba4da94"} Oct 13 09:00:28 crc kubenswrapper[4685]: I1013 09:00:28.646016 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gh7ls" event={"ID":"fb7ad71b-d500-457d-81d4-831ba362a8b4","Type":"ContainerStarted","Data":"8a755736a554fbf6ac52385a58f2fffcfd1809c974a2a49e86643dcac4536d33"} Oct 13 09:00:28 crc kubenswrapper[4685]: I1013 09:00:28.646094 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gh7ls" event={"ID":"fb7ad71b-d500-457d-81d4-831ba362a8b4","Type":"ContainerStarted","Data":"0a246b31e595327c48e1e3c9b103c063f4679c89b27fc9fce6ec0c1475904af5"} Oct 13 09:00:28 crc kubenswrapper[4685]: I1013 09:00:28.647843 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:00:28 crc kubenswrapper[4685]: I1013 09:00:28.648064 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:00:28 crc kubenswrapper[4685]: I1013 09:00:28.700865 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-gh7ls" podStartSLOduration=19.681609964 podStartE2EDuration="24.700845843s" podCreationTimestamp="2025-10-13 09:00:04 +0000 UTC" firstStartedPulling="2025-10-13 09:00:20.531542311 +0000 UTC m=+945.679418072" lastFinishedPulling="2025-10-13 09:00:25.5507782 +0000 UTC m=+950.698653951" observedRunningTime="2025-10-13 09:00:28.69085239 +0000 UTC m=+953.838728161" watchObservedRunningTime="2025-10-13 09:00:28.700845843 +0000 UTC m=+953.848721604" Oct 13 09:00:29 crc kubenswrapper[4685]: I1013 09:00:29.514465 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc" path="/var/lib/kubelet/pods/9df34c78-f8c7-4b0f-9f74-47e1fe50a3fc/volumes" Oct 13 09:00:29 crc kubenswrapper[4685]: I1013 09:00:29.515821 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5" path="/var/lib/kubelet/pods/a54ac13e-48cb-4c0b-8cf9-e7f01e48b8e5/volumes" Oct 13 09:00:29 crc kubenswrapper[4685]: I1013 09:00:29.657092 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" event={"ID":"9ce22efc-b48d-4d42-a01a-31f71c53c3ab","Type":"ContainerStarted","Data":"84ce7f67e60c874cfdcb1883fd997c33a283a503994c815b20be8f0c1ae4a60c"} Oct 13 09:00:29 crc kubenswrapper[4685]: I1013 09:00:29.657347 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" Oct 13 09:00:29 crc kubenswrapper[4685]: I1013 09:00:29.686683 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" podStartSLOduration=4.030167865 podStartE2EDuration="4.686667429s" podCreationTimestamp="2025-10-13 09:00:25 +0000 UTC" firstStartedPulling="2025-10-13 09:00:26.673989639 +0000 UTC m=+951.821865400" lastFinishedPulling="2025-10-13 09:00:27.330489203 +0000 UTC m=+952.478364964" observedRunningTime="2025-10-13 09:00:29.680165341 +0000 UTC m=+954.828041092" watchObservedRunningTime="2025-10-13 09:00:29.686667429 +0000 UTC m=+954.834543190" Oct 13 09:00:31 crc kubenswrapper[4685]: I1013 09:00:31.817260 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 13 09:00:31 crc kubenswrapper[4685]: I1013 09:00:31.898958 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 13 09:00:32 crc kubenswrapper[4685]: I1013 09:00:32.962068 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 13 09:00:34 crc kubenswrapper[4685]: I1013 09:00:34.711226 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-t646z" event={"ID":"85cf8998-48c3-49b9-9a68-d5765b785df4","Type":"ContainerStarted","Data":"3a6d044b329c696dca8ea297b8fb5e6e285339653d9bb655d7ee062e6286f629"} Oct 13 09:00:34 crc kubenswrapper[4685]: I1013 09:00:34.714370 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"7ffc3833-03fd-40ef-b247-4f1c512c2e8b","Type":"ContainerStarted","Data":"2af57fa8f5b2df1c0857fd3c458c7fbe0e0c0c5e8350ef93365f0ceff300dbca"} Oct 13 09:00:34 crc kubenswrapper[4685]: I1013 09:00:34.719091 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2a3319a1-2f17-43cd-9df1-0697ba10aff8","Type":"ContainerStarted","Data":"5473f408533d237ab8d0cda3fe645a5d7ce20fc45745e005dd2cfe8f7316740a"} Oct 13 09:00:34 crc kubenswrapper[4685]: I1013 09:00:34.720744 4685 generic.go:334] "Generic (PLEG): container finished" podID="86735b6c-f0ad-47b7-9af5-f07264a2fbcd" containerID="3e0e62c474576836ee25b947fbfeec0745b5881227fec2e9a29c1ce26e0412a0" exitCode=0 Oct 13 09:00:34 crc kubenswrapper[4685]: I1013 09:00:34.720784 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-dl7rb" event={"ID":"86735b6c-f0ad-47b7-9af5-f07264a2fbcd","Type":"ContainerDied","Data":"3e0e62c474576836ee25b947fbfeec0745b5881227fec2e9a29c1ce26e0412a0"} Oct 13 09:00:34 crc kubenswrapper[4685]: I1013 09:00:34.733737 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-t646z" podStartSLOduration=2.232437038 podStartE2EDuration="9.73371304s" podCreationTimestamp="2025-10-13 09:00:25 +0000 UTC" firstStartedPulling="2025-10-13 09:00:26.585097704 +0000 UTC m=+951.732973465" lastFinishedPulling="2025-10-13 09:00:34.086373716 +0000 UTC m=+959.234249467" observedRunningTime="2025-10-13 09:00:34.727312065 +0000 UTC m=+959.875187826" watchObservedRunningTime="2025-10-13 09:00:34.73371304 +0000 UTC m=+959.881588801" Oct 13 09:00:34 crc kubenswrapper[4685]: I1013 09:00:34.824451 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=17.76320044 podStartE2EDuration="30.824411604s" podCreationTimestamp="2025-10-13 09:00:04 +0000 UTC" firstStartedPulling="2025-10-13 09:00:21.134019265 +0000 UTC m=+946.281895026" lastFinishedPulling="2025-10-13 09:00:34.195230419 +0000 UTC m=+959.343106190" observedRunningTime="2025-10-13 09:00:34.782465585 +0000 UTC m=+959.930341346" watchObservedRunningTime="2025-10-13 09:00:34.824411604 +0000 UTC m=+959.972287355" Oct 13 09:00:34 crc kubenswrapper[4685]: I1013 09:00:34.851831 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=17.580930877 podStartE2EDuration="30.851811145s" podCreationTimestamp="2025-10-13 09:00:04 +0000 UTC" firstStartedPulling="2025-10-13 09:00:20.943511466 +0000 UTC m=+946.091387227" lastFinishedPulling="2025-10-13 09:00:34.214391724 +0000 UTC m=+959.362267495" observedRunningTime="2025-10-13 09:00:34.845097701 +0000 UTC m=+959.992973472" watchObservedRunningTime="2025-10-13 09:00:34.851811145 +0000 UTC m=+959.999686906" Oct 13 09:00:35 crc kubenswrapper[4685]: I1013 09:00:35.700906 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:35 crc kubenswrapper[4685]: I1013 09:00:35.701221 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:35 crc kubenswrapper[4685]: I1013 09:00:35.731562 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-dl7rb" event={"ID":"86735b6c-f0ad-47b7-9af5-f07264a2fbcd","Type":"ContainerStarted","Data":"86f1568c0403f19d5c4ae7d555a2bc831ee4258b00ce001674a8a68d688977ce"} Oct 13 09:00:35 crc kubenswrapper[4685]: I1013 09:00:35.732663 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-dl7rb" Oct 13 09:00:35 crc kubenswrapper[4685]: I1013 09:00:35.754791 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-dl7rb" podStartSLOduration=4.291795762 podStartE2EDuration="10.75476475s" podCreationTimestamp="2025-10-13 09:00:25 +0000 UTC" firstStartedPulling="2025-10-13 09:00:27.254769299 +0000 UTC m=+952.402645060" lastFinishedPulling="2025-10-13 09:00:33.717738277 +0000 UTC m=+958.865614048" observedRunningTime="2025-10-13 09:00:35.751568833 +0000 UTC m=+960.899444604" watchObservedRunningTime="2025-10-13 09:00:35.75476475 +0000 UTC m=+960.902640551" Oct 13 09:00:35 crc kubenswrapper[4685]: I1013 09:00:35.774434 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:36 crc kubenswrapper[4685]: I1013 09:00:36.023502 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:36 crc kubenswrapper[4685]: I1013 09:00:36.023711 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:36 crc kubenswrapper[4685]: I1013 09:00:36.083035 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" Oct 13 09:00:36 crc kubenswrapper[4685]: I1013 09:00:36.087397 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:36 crc kubenswrapper[4685]: I1013 09:00:36.398899 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 13 09:00:36 crc kubenswrapper[4685]: I1013 09:00:36.398960 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 13 09:00:36 crc kubenswrapper[4685]: I1013 09:00:36.451828 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 13 09:00:36 crc kubenswrapper[4685]: I1013 09:00:36.781014 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 13 09:00:36 crc kubenswrapper[4685]: I1013 09:00:36.804180 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 13 09:00:36 crc kubenswrapper[4685]: I1013 09:00:36.901577 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.215713 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.216870 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.218700 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-68272" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.219229 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.219369 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.219810 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.246655 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.386994 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/108cacab-202f-48f0-91f1-664e68d89e28-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"108cacab-202f-48f0-91f1-664e68d89e28\") " pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.387031 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/108cacab-202f-48f0-91f1-664e68d89e28-config\") pod \"ovn-northd-0\" (UID: \"108cacab-202f-48f0-91f1-664e68d89e28\") " pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.387052 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/108cacab-202f-48f0-91f1-664e68d89e28-scripts\") pod \"ovn-northd-0\" (UID: \"108cacab-202f-48f0-91f1-664e68d89e28\") " pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.387112 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/108cacab-202f-48f0-91f1-664e68d89e28-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"108cacab-202f-48f0-91f1-664e68d89e28\") " pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.387208 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/108cacab-202f-48f0-91f1-664e68d89e28-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"108cacab-202f-48f0-91f1-664e68d89e28\") " pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.387280 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljqrl\" (UniqueName: \"kubernetes.io/projected/108cacab-202f-48f0-91f1-664e68d89e28-kube-api-access-ljqrl\") pod \"ovn-northd-0\" (UID: \"108cacab-202f-48f0-91f1-664e68d89e28\") " pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.387352 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/108cacab-202f-48f0-91f1-664e68d89e28-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"108cacab-202f-48f0-91f1-664e68d89e28\") " pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.452684 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-bccdb"] Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.453568 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bccdb" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.467755 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-bccdb"] Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.488481 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/108cacab-202f-48f0-91f1-664e68d89e28-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"108cacab-202f-48f0-91f1-664e68d89e28\") " pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.488525 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/108cacab-202f-48f0-91f1-664e68d89e28-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"108cacab-202f-48f0-91f1-664e68d89e28\") " pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.488546 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljqrl\" (UniqueName: \"kubernetes.io/projected/108cacab-202f-48f0-91f1-664e68d89e28-kube-api-access-ljqrl\") pod \"ovn-northd-0\" (UID: \"108cacab-202f-48f0-91f1-664e68d89e28\") " pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.488573 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/108cacab-202f-48f0-91f1-664e68d89e28-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"108cacab-202f-48f0-91f1-664e68d89e28\") " pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.488624 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/108cacab-202f-48f0-91f1-664e68d89e28-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"108cacab-202f-48f0-91f1-664e68d89e28\") " pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.488644 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/108cacab-202f-48f0-91f1-664e68d89e28-config\") pod \"ovn-northd-0\" (UID: \"108cacab-202f-48f0-91f1-664e68d89e28\") " pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.488662 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/108cacab-202f-48f0-91f1-664e68d89e28-scripts\") pod \"ovn-northd-0\" (UID: \"108cacab-202f-48f0-91f1-664e68d89e28\") " pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.489345 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/108cacab-202f-48f0-91f1-664e68d89e28-scripts\") pod \"ovn-northd-0\" (UID: \"108cacab-202f-48f0-91f1-664e68d89e28\") " pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.489621 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/108cacab-202f-48f0-91f1-664e68d89e28-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"108cacab-202f-48f0-91f1-664e68d89e28\") " pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.490136 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/108cacab-202f-48f0-91f1-664e68d89e28-config\") pod \"ovn-northd-0\" (UID: \"108cacab-202f-48f0-91f1-664e68d89e28\") " pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.494851 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/108cacab-202f-48f0-91f1-664e68d89e28-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"108cacab-202f-48f0-91f1-664e68d89e28\") " pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.495513 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/108cacab-202f-48f0-91f1-664e68d89e28-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"108cacab-202f-48f0-91f1-664e68d89e28\") " pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.498568 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/108cacab-202f-48f0-91f1-664e68d89e28-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"108cacab-202f-48f0-91f1-664e68d89e28\") " pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.507683 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljqrl\" (UniqueName: \"kubernetes.io/projected/108cacab-202f-48f0-91f1-664e68d89e28-kube-api-access-ljqrl\") pod \"ovn-northd-0\" (UID: \"108cacab-202f-48f0-91f1-664e68d89e28\") " pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.541134 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.590048 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cv62\" (UniqueName: \"kubernetes.io/projected/6334858c-1809-484a-a74f-49e28a18e9b2-kube-api-access-6cv62\") pod \"keystone-db-create-bccdb\" (UID: \"6334858c-1809-484a-a74f-49e28a18e9b2\") " pod="openstack/keystone-db-create-bccdb" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.693263 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cv62\" (UniqueName: \"kubernetes.io/projected/6334858c-1809-484a-a74f-49e28a18e9b2-kube-api-access-6cv62\") pod \"keystone-db-create-bccdb\" (UID: \"6334858c-1809-484a-a74f-49e28a18e9b2\") " pod="openstack/keystone-db-create-bccdb" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.731264 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cv62\" (UniqueName: \"kubernetes.io/projected/6334858c-1809-484a-a74f-49e28a18e9b2-kube-api-access-6cv62\") pod \"keystone-db-create-bccdb\" (UID: \"6334858c-1809-484a-a74f-49e28a18e9b2\") " pod="openstack/keystone-db-create-bccdb" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.742997 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-jt65g"] Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.743998 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-jt65g" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.765483 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-jt65g"] Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.765791 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bccdb" Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.850167 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 13 09:00:37 crc kubenswrapper[4685]: W1013 09:00:37.850167 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod108cacab_202f_48f0_91f1_664e68d89e28.slice/crio-918f3daaf2a26a0c464db890ab3399d2551d7cdbad02590569e39bfb85bbd759 WatchSource:0}: Error finding container 918f3daaf2a26a0c464db890ab3399d2551d7cdbad02590569e39bfb85bbd759: Status 404 returned error can't find the container with id 918f3daaf2a26a0c464db890ab3399d2551d7cdbad02590569e39bfb85bbd759 Oct 13 09:00:37 crc kubenswrapper[4685]: I1013 09:00:37.916445 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-928xt\" (UniqueName: \"kubernetes.io/projected/a48e241f-11b8-4c11-a17a-0d96b2b37df2-kube-api-access-928xt\") pod \"placement-db-create-jt65g\" (UID: \"a48e241f-11b8-4c11-a17a-0d96b2b37df2\") " pod="openstack/placement-db-create-jt65g" Oct 13 09:00:38 crc kubenswrapper[4685]: I1013 09:00:38.017760 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-928xt\" (UniqueName: \"kubernetes.io/projected/a48e241f-11b8-4c11-a17a-0d96b2b37df2-kube-api-access-928xt\") pod \"placement-db-create-jt65g\" (UID: \"a48e241f-11b8-4c11-a17a-0d96b2b37df2\") " pod="openstack/placement-db-create-jt65g" Oct 13 09:00:38 crc kubenswrapper[4685]: I1013 09:00:38.036097 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-928xt\" (UniqueName: \"kubernetes.io/projected/a48e241f-11b8-4c11-a17a-0d96b2b37df2-kube-api-access-928xt\") pod \"placement-db-create-jt65g\" (UID: \"a48e241f-11b8-4c11-a17a-0d96b2b37df2\") " pod="openstack/placement-db-create-jt65g" Oct 13 09:00:38 crc kubenswrapper[4685]: I1013 09:00:38.084802 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-jt65g" Oct 13 09:00:38 crc kubenswrapper[4685]: I1013 09:00:38.261576 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-bccdb"] Oct 13 09:00:38 crc kubenswrapper[4685]: W1013 09:00:38.263710 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6334858c_1809_484a_a74f_49e28a18e9b2.slice/crio-ed00bb0ad2fe035df5a77a5fd28a7df01be67cea6cc2f5faf14e8dfa37767059 WatchSource:0}: Error finding container ed00bb0ad2fe035df5a77a5fd28a7df01be67cea6cc2f5faf14e8dfa37767059: Status 404 returned error can't find the container with id ed00bb0ad2fe035df5a77a5fd28a7df01be67cea6cc2f5faf14e8dfa37767059 Oct 13 09:00:38 crc kubenswrapper[4685]: I1013 09:00:38.500476 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-jt65g"] Oct 13 09:00:38 crc kubenswrapper[4685]: W1013 09:00:38.513027 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda48e241f_11b8_4c11_a17a_0d96b2b37df2.slice/crio-6384ab1f810b28d7b5956d613cee412371d90a4fe43c4f63d7ab3327f71ae2d9 WatchSource:0}: Error finding container 6384ab1f810b28d7b5956d613cee412371d90a4fe43c4f63d7ab3327f71ae2d9: Status 404 returned error can't find the container with id 6384ab1f810b28d7b5956d613cee412371d90a4fe43c4f63d7ab3327f71ae2d9 Oct 13 09:00:38 crc kubenswrapper[4685]: I1013 09:00:38.761598 4685 generic.go:334] "Generic (PLEG): container finished" podID="6334858c-1809-484a-a74f-49e28a18e9b2" containerID="13ac4790a8c78f7ce3c83e12f382c943e0ce69f5e6c37d3f5a09d4d29b11b351" exitCode=0 Oct 13 09:00:38 crc kubenswrapper[4685]: I1013 09:00:38.761663 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bccdb" event={"ID":"6334858c-1809-484a-a74f-49e28a18e9b2","Type":"ContainerDied","Data":"13ac4790a8c78f7ce3c83e12f382c943e0ce69f5e6c37d3f5a09d4d29b11b351"} Oct 13 09:00:38 crc kubenswrapper[4685]: I1013 09:00:38.761685 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bccdb" event={"ID":"6334858c-1809-484a-a74f-49e28a18e9b2","Type":"ContainerStarted","Data":"ed00bb0ad2fe035df5a77a5fd28a7df01be67cea6cc2f5faf14e8dfa37767059"} Oct 13 09:00:38 crc kubenswrapper[4685]: I1013 09:00:38.762796 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"108cacab-202f-48f0-91f1-664e68d89e28","Type":"ContainerStarted","Data":"918f3daaf2a26a0c464db890ab3399d2551d7cdbad02590569e39bfb85bbd759"} Oct 13 09:00:38 crc kubenswrapper[4685]: I1013 09:00:38.765757 4685 generic.go:334] "Generic (PLEG): container finished" podID="a48e241f-11b8-4c11-a17a-0d96b2b37df2" containerID="24b54200b33177414f9bf29827091c56f176ceea6d18f5f169c572ed2d0be71f" exitCode=0 Oct 13 09:00:38 crc kubenswrapper[4685]: I1013 09:00:38.765881 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-jt65g" event={"ID":"a48e241f-11b8-4c11-a17a-0d96b2b37df2","Type":"ContainerDied","Data":"24b54200b33177414f9bf29827091c56f176ceea6d18f5f169c572ed2d0be71f"} Oct 13 09:00:38 crc kubenswrapper[4685]: I1013 09:00:38.766008 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-jt65g" event={"ID":"a48e241f-11b8-4c11-a17a-0d96b2b37df2","Type":"ContainerStarted","Data":"6384ab1f810b28d7b5956d613cee412371d90a4fe43c4f63d7ab3327f71ae2d9"} Oct 13 09:00:39 crc kubenswrapper[4685]: I1013 09:00:39.658349 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 13 09:00:39 crc kubenswrapper[4685]: I1013 09:00:39.756606 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-dl7rb"] Oct 13 09:00:39 crc kubenswrapper[4685]: I1013 09:00:39.757105 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-dl7rb" podUID="86735b6c-f0ad-47b7-9af5-f07264a2fbcd" containerName="dnsmasq-dns" containerID="cri-o://86f1568c0403f19d5c4ae7d555a2bc831ee4258b00ce001674a8a68d688977ce" gracePeriod=10 Oct 13 09:00:39 crc kubenswrapper[4685]: I1013 09:00:39.808678 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-7fttg"] Oct 13 09:00:39 crc kubenswrapper[4685]: I1013 09:00:39.810263 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" Oct 13 09:00:39 crc kubenswrapper[4685]: I1013 09:00:39.825237 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-7fttg"] Oct 13 09:00:39 crc kubenswrapper[4685]: I1013 09:00:39.972253 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-7fttg\" (UID: \"13c3d843-ad90-41fc-9f5f-6935eb976489\") " pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" Oct 13 09:00:39 crc kubenswrapper[4685]: I1013 09:00:39.972578 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-7fttg\" (UID: \"13c3d843-ad90-41fc-9f5f-6935eb976489\") " pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" Oct 13 09:00:39 crc kubenswrapper[4685]: I1013 09:00:39.972634 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-7fttg\" (UID: \"13c3d843-ad90-41fc-9f5f-6935eb976489\") " pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" Oct 13 09:00:39 crc kubenswrapper[4685]: I1013 09:00:39.972691 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-config\") pod \"dnsmasq-dns-b8fbc5445-7fttg\" (UID: \"13c3d843-ad90-41fc-9f5f-6935eb976489\") " pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" Oct 13 09:00:39 crc kubenswrapper[4685]: I1013 09:00:39.972720 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c65ws\" (UniqueName: \"kubernetes.io/projected/13c3d843-ad90-41fc-9f5f-6935eb976489-kube-api-access-c65ws\") pod \"dnsmasq-dns-b8fbc5445-7fttg\" (UID: \"13c3d843-ad90-41fc-9f5f-6935eb976489\") " pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.074305 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-config\") pod \"dnsmasq-dns-b8fbc5445-7fttg\" (UID: \"13c3d843-ad90-41fc-9f5f-6935eb976489\") " pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.075545 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c65ws\" (UniqueName: \"kubernetes.io/projected/13c3d843-ad90-41fc-9f5f-6935eb976489-kube-api-access-c65ws\") pod \"dnsmasq-dns-b8fbc5445-7fttg\" (UID: \"13c3d843-ad90-41fc-9f5f-6935eb976489\") " pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.075673 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-7fttg\" (UID: \"13c3d843-ad90-41fc-9f5f-6935eb976489\") " pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.076083 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-7fttg\" (UID: \"13c3d843-ad90-41fc-9f5f-6935eb976489\") " pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.076218 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-7fttg\" (UID: \"13c3d843-ad90-41fc-9f5f-6935eb976489\") " pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.075479 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-config\") pod \"dnsmasq-dns-b8fbc5445-7fttg\" (UID: \"13c3d843-ad90-41fc-9f5f-6935eb976489\") " pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.076666 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-7fttg\" (UID: \"13c3d843-ad90-41fc-9f5f-6935eb976489\") " pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.076782 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-7fttg\" (UID: \"13c3d843-ad90-41fc-9f5f-6935eb976489\") " pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.077193 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-7fttg\" (UID: \"13c3d843-ad90-41fc-9f5f-6935eb976489\") " pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.110065 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c65ws\" (UniqueName: \"kubernetes.io/projected/13c3d843-ad90-41fc-9f5f-6935eb976489-kube-api-access-c65ws\") pod \"dnsmasq-dns-b8fbc5445-7fttg\" (UID: \"13c3d843-ad90-41fc-9f5f-6935eb976489\") " pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.226436 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.289541 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bccdb" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.293194 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-dl7rb" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.310556 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-jt65g" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.384044 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-dns-svc\") pod \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\" (UID: \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\") " Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.384097 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-ovsdbserver-sb\") pod \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\" (UID: \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\") " Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.384136 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-ovsdbserver-nb\") pod \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\" (UID: \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\") " Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.384172 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-config\") pod \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\" (UID: \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\") " Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.384283 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x25b2\" (UniqueName: \"kubernetes.io/projected/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-kube-api-access-x25b2\") pod \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\" (UID: \"86735b6c-f0ad-47b7-9af5-f07264a2fbcd\") " Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.384315 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-928xt\" (UniqueName: \"kubernetes.io/projected/a48e241f-11b8-4c11-a17a-0d96b2b37df2-kube-api-access-928xt\") pod \"a48e241f-11b8-4c11-a17a-0d96b2b37df2\" (UID: \"a48e241f-11b8-4c11-a17a-0d96b2b37df2\") " Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.384338 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cv62\" (UniqueName: \"kubernetes.io/projected/6334858c-1809-484a-a74f-49e28a18e9b2-kube-api-access-6cv62\") pod \"6334858c-1809-484a-a74f-49e28a18e9b2\" (UID: \"6334858c-1809-484a-a74f-49e28a18e9b2\") " Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.391851 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-kube-api-access-x25b2" (OuterVolumeSpecName: "kube-api-access-x25b2") pod "86735b6c-f0ad-47b7-9af5-f07264a2fbcd" (UID: "86735b6c-f0ad-47b7-9af5-f07264a2fbcd"). InnerVolumeSpecName "kube-api-access-x25b2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.392530 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6334858c-1809-484a-a74f-49e28a18e9b2-kube-api-access-6cv62" (OuterVolumeSpecName: "kube-api-access-6cv62") pod "6334858c-1809-484a-a74f-49e28a18e9b2" (UID: "6334858c-1809-484a-a74f-49e28a18e9b2"). InnerVolumeSpecName "kube-api-access-6cv62". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.394099 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a48e241f-11b8-4c11-a17a-0d96b2b37df2-kube-api-access-928xt" (OuterVolumeSpecName: "kube-api-access-928xt") pod "a48e241f-11b8-4c11-a17a-0d96b2b37df2" (UID: "a48e241f-11b8-4c11-a17a-0d96b2b37df2"). InnerVolumeSpecName "kube-api-access-928xt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.429732 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "86735b6c-f0ad-47b7-9af5-f07264a2fbcd" (UID: "86735b6c-f0ad-47b7-9af5-f07264a2fbcd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.466292 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "86735b6c-f0ad-47b7-9af5-f07264a2fbcd" (UID: "86735b6c-f0ad-47b7-9af5-f07264a2fbcd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.469824 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-config" (OuterVolumeSpecName: "config") pod "86735b6c-f0ad-47b7-9af5-f07264a2fbcd" (UID: "86735b6c-f0ad-47b7-9af5-f07264a2fbcd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.480646 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "86735b6c-f0ad-47b7-9af5-f07264a2fbcd" (UID: "86735b6c-f0ad-47b7-9af5-f07264a2fbcd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.486960 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-config\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.487038 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x25b2\" (UniqueName: \"kubernetes.io/projected/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-kube-api-access-x25b2\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.487048 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-928xt\" (UniqueName: \"kubernetes.io/projected/a48e241f-11b8-4c11-a17a-0d96b2b37df2-kube-api-access-928xt\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.487077 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cv62\" (UniqueName: \"kubernetes.io/projected/6334858c-1809-484a-a74f-49e28a18e9b2-kube-api-access-6cv62\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.487087 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.487096 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.487106 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86735b6c-f0ad-47b7-9af5-f07264a2fbcd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.730500 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-7fttg"] Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.791233 4685 generic.go:334] "Generic (PLEG): container finished" podID="86735b6c-f0ad-47b7-9af5-f07264a2fbcd" containerID="86f1568c0403f19d5c4ae7d555a2bc831ee4258b00ce001674a8a68d688977ce" exitCode=0 Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.791320 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-dl7rb" event={"ID":"86735b6c-f0ad-47b7-9af5-f07264a2fbcd","Type":"ContainerDied","Data":"86f1568c0403f19d5c4ae7d555a2bc831ee4258b00ce001674a8a68d688977ce"} Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.791350 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-dl7rb" event={"ID":"86735b6c-f0ad-47b7-9af5-f07264a2fbcd","Type":"ContainerDied","Data":"3a7920f25591dd079e23c2fd29360fb24d151297f61d242c53353415e0ea47f0"} Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.791371 4685 scope.go:117] "RemoveContainer" containerID="86f1568c0403f19d5c4ae7d555a2bc831ee4258b00ce001674a8a68d688977ce" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.791503 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-dl7rb" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.801365 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-jt65g" event={"ID":"a48e241f-11b8-4c11-a17a-0d96b2b37df2","Type":"ContainerDied","Data":"6384ab1f810b28d7b5956d613cee412371d90a4fe43c4f63d7ab3327f71ae2d9"} Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.801411 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6384ab1f810b28d7b5956d613cee412371d90a4fe43c4f63d7ab3327f71ae2d9" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.801471 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-jt65g" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.805646 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bccdb" event={"ID":"6334858c-1809-484a-a74f-49e28a18e9b2","Type":"ContainerDied","Data":"ed00bb0ad2fe035df5a77a5fd28a7df01be67cea6cc2f5faf14e8dfa37767059"} Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.805687 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed00bb0ad2fe035df5a77a5fd28a7df01be67cea6cc2f5faf14e8dfa37767059" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.805746 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bccdb" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.836951 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-dl7rb"] Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.844272 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-dl7rb"] Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.853549 4685 scope.go:117] "RemoveContainer" containerID="3e0e62c474576836ee25b947fbfeec0745b5881227fec2e9a29c1ce26e0412a0" Oct 13 09:00:40 crc kubenswrapper[4685]: W1013 09:00:40.854076 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13c3d843_ad90_41fc_9f5f_6935eb976489.slice/crio-684d8eb91a1de29de2393f631e657a8894b63dbdc6da4fc5914ff63d1f4e86b4 WatchSource:0}: Error finding container 684d8eb91a1de29de2393f631e657a8894b63dbdc6da4fc5914ff63d1f4e86b4: Status 404 returned error can't find the container with id 684d8eb91a1de29de2393f631e657a8894b63dbdc6da4fc5914ff63d1f4e86b4 Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.877862 4685 scope.go:117] "RemoveContainer" containerID="86f1568c0403f19d5c4ae7d555a2bc831ee4258b00ce001674a8a68d688977ce" Oct 13 09:00:40 crc kubenswrapper[4685]: E1013 09:00:40.878316 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86f1568c0403f19d5c4ae7d555a2bc831ee4258b00ce001674a8a68d688977ce\": container with ID starting with 86f1568c0403f19d5c4ae7d555a2bc831ee4258b00ce001674a8a68d688977ce not found: ID does not exist" containerID="86f1568c0403f19d5c4ae7d555a2bc831ee4258b00ce001674a8a68d688977ce" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.878362 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86f1568c0403f19d5c4ae7d555a2bc831ee4258b00ce001674a8a68d688977ce"} err="failed to get container status \"86f1568c0403f19d5c4ae7d555a2bc831ee4258b00ce001674a8a68d688977ce\": rpc error: code = NotFound desc = could not find container \"86f1568c0403f19d5c4ae7d555a2bc831ee4258b00ce001674a8a68d688977ce\": container with ID starting with 86f1568c0403f19d5c4ae7d555a2bc831ee4258b00ce001674a8a68d688977ce not found: ID does not exist" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.878384 4685 scope.go:117] "RemoveContainer" containerID="3e0e62c474576836ee25b947fbfeec0745b5881227fec2e9a29c1ce26e0412a0" Oct 13 09:00:40 crc kubenswrapper[4685]: E1013 09:00:40.878818 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e0e62c474576836ee25b947fbfeec0745b5881227fec2e9a29c1ce26e0412a0\": container with ID starting with 3e0e62c474576836ee25b947fbfeec0745b5881227fec2e9a29c1ce26e0412a0 not found: ID does not exist" containerID="3e0e62c474576836ee25b947fbfeec0745b5881227fec2e9a29c1ce26e0412a0" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.878847 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e0e62c474576836ee25b947fbfeec0745b5881227fec2e9a29c1ce26e0412a0"} err="failed to get container status \"3e0e62c474576836ee25b947fbfeec0745b5881227fec2e9a29c1ce26e0412a0\": rpc error: code = NotFound desc = could not find container \"3e0e62c474576836ee25b947fbfeec0745b5881227fec2e9a29c1ce26e0412a0\": container with ID starting with 3e0e62c474576836ee25b947fbfeec0745b5881227fec2e9a29c1ce26e0412a0 not found: ID does not exist" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.952214 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 13 09:00:40 crc kubenswrapper[4685]: E1013 09:00:40.953206 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6334858c-1809-484a-a74f-49e28a18e9b2" containerName="mariadb-database-create" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.953223 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6334858c-1809-484a-a74f-49e28a18e9b2" containerName="mariadb-database-create" Oct 13 09:00:40 crc kubenswrapper[4685]: E1013 09:00:40.953249 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a48e241f-11b8-4c11-a17a-0d96b2b37df2" containerName="mariadb-database-create" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.953258 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="a48e241f-11b8-4c11-a17a-0d96b2b37df2" containerName="mariadb-database-create" Oct 13 09:00:40 crc kubenswrapper[4685]: E1013 09:00:40.953274 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86735b6c-f0ad-47b7-9af5-f07264a2fbcd" containerName="dnsmasq-dns" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.953282 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="86735b6c-f0ad-47b7-9af5-f07264a2fbcd" containerName="dnsmasq-dns" Oct 13 09:00:40 crc kubenswrapper[4685]: E1013 09:00:40.953303 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86735b6c-f0ad-47b7-9af5-f07264a2fbcd" containerName="init" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.953311 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="86735b6c-f0ad-47b7-9af5-f07264a2fbcd" containerName="init" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.953493 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="6334858c-1809-484a-a74f-49e28a18e9b2" containerName="mariadb-database-create" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.953515 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="a48e241f-11b8-4c11-a17a-0d96b2b37df2" containerName="mariadb-database-create" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.953531 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="86735b6c-f0ad-47b7-9af5-f07264a2fbcd" containerName="dnsmasq-dns" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.969960 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.970873 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.978330 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.978638 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.978854 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-xhnh9" Oct 13 09:00:40 crc kubenswrapper[4685]: I1013 09:00:40.979082 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.099381 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0791156b-11f2-43ad-b910-3e42b4d6670f-cache\") pod \"swift-storage-0\" (UID: \"0791156b-11f2-43ad-b910-3e42b4d6670f\") " pod="openstack/swift-storage-0" Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.099426 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrrrn\" (UniqueName: \"kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-kube-api-access-qrrrn\") pod \"swift-storage-0\" (UID: \"0791156b-11f2-43ad-b910-3e42b4d6670f\") " pod="openstack/swift-storage-0" Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.099465 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0791156b-11f2-43ad-b910-3e42b4d6670f-lock\") pod \"swift-storage-0\" (UID: \"0791156b-11f2-43ad-b910-3e42b4d6670f\") " pod="openstack/swift-storage-0" Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.099507 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"0791156b-11f2-43ad-b910-3e42b4d6670f\") " pod="openstack/swift-storage-0" Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.099522 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-etc-swift\") pod \"swift-storage-0\" (UID: \"0791156b-11f2-43ad-b910-3e42b4d6670f\") " pod="openstack/swift-storage-0" Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.200697 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0791156b-11f2-43ad-b910-3e42b4d6670f-lock\") pod \"swift-storage-0\" (UID: \"0791156b-11f2-43ad-b910-3e42b4d6670f\") " pod="openstack/swift-storage-0" Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.200755 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"0791156b-11f2-43ad-b910-3e42b4d6670f\") " pod="openstack/swift-storage-0" Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.200777 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-etc-swift\") pod \"swift-storage-0\" (UID: \"0791156b-11f2-43ad-b910-3e42b4d6670f\") " pod="openstack/swift-storage-0" Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.200848 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0791156b-11f2-43ad-b910-3e42b4d6670f-cache\") pod \"swift-storage-0\" (UID: \"0791156b-11f2-43ad-b910-3e42b4d6670f\") " pod="openstack/swift-storage-0" Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.200878 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrrrn\" (UniqueName: \"kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-kube-api-access-qrrrn\") pod \"swift-storage-0\" (UID: \"0791156b-11f2-43ad-b910-3e42b4d6670f\") " pod="openstack/swift-storage-0" Oct 13 09:00:41 crc kubenswrapper[4685]: E1013 09:00:41.201314 4685 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 13 09:00:41 crc kubenswrapper[4685]: E1013 09:00:41.201396 4685 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.201444 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0791156b-11f2-43ad-b910-3e42b4d6670f-cache\") pod \"swift-storage-0\" (UID: \"0791156b-11f2-43ad-b910-3e42b4d6670f\") " pod="openstack/swift-storage-0" Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.201487 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"0791156b-11f2-43ad-b910-3e42b4d6670f\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/swift-storage-0" Oct 13 09:00:41 crc kubenswrapper[4685]: E1013 09:00:41.201660 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-etc-swift podName:0791156b-11f2-43ad-b910-3e42b4d6670f nodeName:}" failed. No retries permitted until 2025-10-13 09:00:41.701600313 +0000 UTC m=+966.849476074 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-etc-swift") pod "swift-storage-0" (UID: "0791156b-11f2-43ad-b910-3e42b4d6670f") : configmap "swift-ring-files" not found Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.206057 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0791156b-11f2-43ad-b910-3e42b4d6670f-lock\") pod \"swift-storage-0\" (UID: \"0791156b-11f2-43ad-b910-3e42b4d6670f\") " pod="openstack/swift-storage-0" Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.219744 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrrrn\" (UniqueName: \"kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-kube-api-access-qrrrn\") pod \"swift-storage-0\" (UID: \"0791156b-11f2-43ad-b910-3e42b4d6670f\") " pod="openstack/swift-storage-0" Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.233761 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"0791156b-11f2-43ad-b910-3e42b4d6670f\") " pod="openstack/swift-storage-0" Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.519538 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86735b6c-f0ad-47b7-9af5-f07264a2fbcd" path="/var/lib/kubelet/pods/86735b6c-f0ad-47b7-9af5-f07264a2fbcd/volumes" Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.709380 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-etc-swift\") pod \"swift-storage-0\" (UID: \"0791156b-11f2-43ad-b910-3e42b4d6670f\") " pod="openstack/swift-storage-0" Oct 13 09:00:41 crc kubenswrapper[4685]: E1013 09:00:41.709556 4685 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 13 09:00:41 crc kubenswrapper[4685]: E1013 09:00:41.709731 4685 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 13 09:00:41 crc kubenswrapper[4685]: E1013 09:00:41.709794 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-etc-swift podName:0791156b-11f2-43ad-b910-3e42b4d6670f nodeName:}" failed. No retries permitted until 2025-10-13 09:00:42.709771844 +0000 UTC m=+967.857647595 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-etc-swift") pod "swift-storage-0" (UID: "0791156b-11f2-43ad-b910-3e42b4d6670f") : configmap "swift-ring-files" not found Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.816157 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"108cacab-202f-48f0-91f1-664e68d89e28","Type":"ContainerStarted","Data":"8db923c74815e44f7770af4e8adbe69d9ae4ffab8710a638f3aec4c9dc10941a"} Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.817582 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"108cacab-202f-48f0-91f1-664e68d89e28","Type":"ContainerStarted","Data":"9cdf7eda0aaeeb06c38ff65cc3a6e8f3bceb5a5fb46ec641c2a18bac60b89a92"} Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.818537 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.821281 4685 generic.go:334] "Generic (PLEG): container finished" podID="13c3d843-ad90-41fc-9f5f-6935eb976489" containerID="c834de30ea8b9cd3a8cb84f6f316c5188d2ff326b7a4d20aa5123e9f211bea84" exitCode=0 Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.821315 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" event={"ID":"13c3d843-ad90-41fc-9f5f-6935eb976489","Type":"ContainerDied","Data":"c834de30ea8b9cd3a8cb84f6f316c5188d2ff326b7a4d20aa5123e9f211bea84"} Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.821335 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" event={"ID":"13c3d843-ad90-41fc-9f5f-6935eb976489","Type":"ContainerStarted","Data":"684d8eb91a1de29de2393f631e657a8894b63dbdc6da4fc5914ff63d1f4e86b4"} Oct 13 09:00:41 crc kubenswrapper[4685]: I1013 09:00:41.848752 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.7697528550000001 podStartE2EDuration="4.848726261s" podCreationTimestamp="2025-10-13 09:00:37 +0000 UTC" firstStartedPulling="2025-10-13 09:00:37.852825395 +0000 UTC m=+963.000701156" lastFinishedPulling="2025-10-13 09:00:40.931798801 +0000 UTC m=+966.079674562" observedRunningTime="2025-10-13 09:00:41.845514353 +0000 UTC m=+966.993390144" watchObservedRunningTime="2025-10-13 09:00:41.848726261 +0000 UTC m=+966.996602042" Oct 13 09:00:42 crc kubenswrapper[4685]: I1013 09:00:42.725743 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-etc-swift\") pod \"swift-storage-0\" (UID: \"0791156b-11f2-43ad-b910-3e42b4d6670f\") " pod="openstack/swift-storage-0" Oct 13 09:00:42 crc kubenswrapper[4685]: E1013 09:00:42.725902 4685 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 13 09:00:42 crc kubenswrapper[4685]: E1013 09:00:42.726246 4685 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 13 09:00:42 crc kubenswrapper[4685]: E1013 09:00:42.726317 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-etc-swift podName:0791156b-11f2-43ad-b910-3e42b4d6670f nodeName:}" failed. No retries permitted until 2025-10-13 09:00:44.726293201 +0000 UTC m=+969.874168962 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-etc-swift") pod "swift-storage-0" (UID: "0791156b-11f2-43ad-b910-3e42b4d6670f") : configmap "swift-ring-files" not found Oct 13 09:00:42 crc kubenswrapper[4685]: I1013 09:00:42.830587 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" event={"ID":"13c3d843-ad90-41fc-9f5f-6935eb976489","Type":"ContainerStarted","Data":"b817b4e33a9c45dbe7eac5187fcb05112d559a3966f81e8d5ce20dbe4d35211f"} Oct 13 09:00:42 crc kubenswrapper[4685]: I1013 09:00:42.863521 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" podStartSLOduration=3.8634983800000002 podStartE2EDuration="3.86349838s" podCreationTimestamp="2025-10-13 09:00:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:00:42.862900933 +0000 UTC m=+968.010776704" watchObservedRunningTime="2025-10-13 09:00:42.86349838 +0000 UTC m=+968.011374141" Oct 13 09:00:43 crc kubenswrapper[4685]: I1013 09:00:43.101106 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-r4l79"] Oct 13 09:00:43 crc kubenswrapper[4685]: I1013 09:00:43.102317 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-r4l79" Oct 13 09:00:43 crc kubenswrapper[4685]: I1013 09:00:43.111354 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-r4l79"] Oct 13 09:00:43 crc kubenswrapper[4685]: I1013 09:00:43.233635 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvrkd\" (UniqueName: \"kubernetes.io/projected/86078d92-4584-4ed5-9d3b-280e6678e18d-kube-api-access-wvrkd\") pod \"glance-db-create-r4l79\" (UID: \"86078d92-4584-4ed5-9d3b-280e6678e18d\") " pod="openstack/glance-db-create-r4l79" Oct 13 09:00:43 crc kubenswrapper[4685]: I1013 09:00:43.334848 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvrkd\" (UniqueName: \"kubernetes.io/projected/86078d92-4584-4ed5-9d3b-280e6678e18d-kube-api-access-wvrkd\") pod \"glance-db-create-r4l79\" (UID: \"86078d92-4584-4ed5-9d3b-280e6678e18d\") " pod="openstack/glance-db-create-r4l79" Oct 13 09:00:43 crc kubenswrapper[4685]: I1013 09:00:43.379077 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvrkd\" (UniqueName: \"kubernetes.io/projected/86078d92-4584-4ed5-9d3b-280e6678e18d-kube-api-access-wvrkd\") pod \"glance-db-create-r4l79\" (UID: \"86078d92-4584-4ed5-9d3b-280e6678e18d\") " pod="openstack/glance-db-create-r4l79" Oct 13 09:00:43 crc kubenswrapper[4685]: I1013 09:00:43.424632 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-r4l79" Oct 13 09:00:43 crc kubenswrapper[4685]: I1013 09:00:43.836824 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" Oct 13 09:00:43 crc kubenswrapper[4685]: I1013 09:00:43.859891 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-r4l79"] Oct 13 09:00:43 crc kubenswrapper[4685]: W1013 09:00:43.864408 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86078d92_4584_4ed5_9d3b_280e6678e18d.slice/crio-f783cf1eee0c13170cfd5e5113807b7df9cfbf4b70b436dae18476c2a07657af WatchSource:0}: Error finding container f783cf1eee0c13170cfd5e5113807b7df9cfbf4b70b436dae18476c2a07657af: Status 404 returned error can't find the container with id f783cf1eee0c13170cfd5e5113807b7df9cfbf4b70b436dae18476c2a07657af Oct 13 09:00:44 crc kubenswrapper[4685]: I1013 09:00:44.756093 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-etc-swift\") pod \"swift-storage-0\" (UID: \"0791156b-11f2-43ad-b910-3e42b4d6670f\") " pod="openstack/swift-storage-0" Oct 13 09:00:44 crc kubenswrapper[4685]: E1013 09:00:44.756340 4685 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 13 09:00:44 crc kubenswrapper[4685]: E1013 09:00:44.756475 4685 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 13 09:00:44 crc kubenswrapper[4685]: E1013 09:00:44.756536 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-etc-swift podName:0791156b-11f2-43ad-b910-3e42b4d6670f nodeName:}" failed. No retries permitted until 2025-10-13 09:00:48.756514878 +0000 UTC m=+973.904390659 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-etc-swift") pod "swift-storage-0" (UID: "0791156b-11f2-43ad-b910-3e42b4d6670f") : configmap "swift-ring-files" not found Oct 13 09:00:44 crc kubenswrapper[4685]: I1013 09:00:44.860038 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-2qrrt"] Oct 13 09:00:44 crc kubenswrapper[4685]: I1013 09:00:44.870696 4685 generic.go:334] "Generic (PLEG): container finished" podID="86078d92-4584-4ed5-9d3b-280e6678e18d" containerID="48a7d3f55efb9dfe5ba19f8f906ecb1eb6aa0bb2710440cf37b87eabc8bd6568" exitCode=0 Oct 13 09:00:44 crc kubenswrapper[4685]: I1013 09:00:44.876855 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-r4l79" event={"ID":"86078d92-4584-4ed5-9d3b-280e6678e18d","Type":"ContainerDied","Data":"48a7d3f55efb9dfe5ba19f8f906ecb1eb6aa0bb2710440cf37b87eabc8bd6568"} Oct 13 09:00:44 crc kubenswrapper[4685]: I1013 09:00:44.876905 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-r4l79" event={"ID":"86078d92-4584-4ed5-9d3b-280e6678e18d","Type":"ContainerStarted","Data":"f783cf1eee0c13170cfd5e5113807b7df9cfbf4b70b436dae18476c2a07657af"} Oct 13 09:00:44 crc kubenswrapper[4685]: I1013 09:00:44.877012 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:44 crc kubenswrapper[4685]: I1013 09:00:44.879559 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 13 09:00:44 crc kubenswrapper[4685]: I1013 09:00:44.880176 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 13 09:00:44 crc kubenswrapper[4685]: I1013 09:00:44.880362 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 13 09:00:44 crc kubenswrapper[4685]: I1013 09:00:44.880459 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-2qrrt"] Oct 13 09:00:44 crc kubenswrapper[4685]: I1013 09:00:44.964193 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-etc-swift\") pod \"swift-ring-rebalance-2qrrt\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:44 crc kubenswrapper[4685]: I1013 09:00:44.964277 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5tv7\" (UniqueName: \"kubernetes.io/projected/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-kube-api-access-h5tv7\") pod \"swift-ring-rebalance-2qrrt\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:44 crc kubenswrapper[4685]: I1013 09:00:44.964310 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-ring-data-devices\") pod \"swift-ring-rebalance-2qrrt\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:44 crc kubenswrapper[4685]: I1013 09:00:44.964379 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-dispersionconf\") pod \"swift-ring-rebalance-2qrrt\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:44 crc kubenswrapper[4685]: I1013 09:00:44.964446 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-scripts\") pod \"swift-ring-rebalance-2qrrt\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:44 crc kubenswrapper[4685]: I1013 09:00:44.964489 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-combined-ca-bundle\") pod \"swift-ring-rebalance-2qrrt\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:44 crc kubenswrapper[4685]: I1013 09:00:44.964545 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-swiftconf\") pod \"swift-ring-rebalance-2qrrt\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:45 crc kubenswrapper[4685]: I1013 09:00:45.066010 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-etc-swift\") pod \"swift-ring-rebalance-2qrrt\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:45 crc kubenswrapper[4685]: I1013 09:00:45.066086 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5tv7\" (UniqueName: \"kubernetes.io/projected/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-kube-api-access-h5tv7\") pod \"swift-ring-rebalance-2qrrt\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:45 crc kubenswrapper[4685]: I1013 09:00:45.066120 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-ring-data-devices\") pod \"swift-ring-rebalance-2qrrt\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:45 crc kubenswrapper[4685]: I1013 09:00:45.066170 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-dispersionconf\") pod \"swift-ring-rebalance-2qrrt\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:45 crc kubenswrapper[4685]: I1013 09:00:45.066214 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-scripts\") pod \"swift-ring-rebalance-2qrrt\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:45 crc kubenswrapper[4685]: I1013 09:00:45.066244 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-combined-ca-bundle\") pod \"swift-ring-rebalance-2qrrt\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:45 crc kubenswrapper[4685]: I1013 09:00:45.066300 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-swiftconf\") pod \"swift-ring-rebalance-2qrrt\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:45 crc kubenswrapper[4685]: I1013 09:00:45.066889 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-scripts\") pod \"swift-ring-rebalance-2qrrt\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:45 crc kubenswrapper[4685]: I1013 09:00:45.067061 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-ring-data-devices\") pod \"swift-ring-rebalance-2qrrt\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:45 crc kubenswrapper[4685]: I1013 09:00:45.067368 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-etc-swift\") pod \"swift-ring-rebalance-2qrrt\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:45 crc kubenswrapper[4685]: I1013 09:00:45.071970 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-swiftconf\") pod \"swift-ring-rebalance-2qrrt\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:45 crc kubenswrapper[4685]: I1013 09:00:45.074577 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-combined-ca-bundle\") pod \"swift-ring-rebalance-2qrrt\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:45 crc kubenswrapper[4685]: I1013 09:00:45.077513 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-dispersionconf\") pod \"swift-ring-rebalance-2qrrt\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:45 crc kubenswrapper[4685]: I1013 09:00:45.084553 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5tv7\" (UniqueName: \"kubernetes.io/projected/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-kube-api-access-h5tv7\") pod \"swift-ring-rebalance-2qrrt\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:45 crc kubenswrapper[4685]: I1013 09:00:45.199979 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:00:45 crc kubenswrapper[4685]: I1013 09:00:45.680440 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-2qrrt"] Oct 13 09:00:45 crc kubenswrapper[4685]: I1013 09:00:45.881705 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-2qrrt" event={"ID":"b884b6de-f048-4f14-b8b0-4775fa1d4cd1","Type":"ContainerStarted","Data":"ac18b068f1de112c3337f0ccad2c80b24f38a378ed6422cdec7f5e9bf6e711ce"} Oct 13 09:00:46 crc kubenswrapper[4685]: I1013 09:00:46.221223 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-r4l79" Oct 13 09:00:46 crc kubenswrapper[4685]: I1013 09:00:46.285451 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvrkd\" (UniqueName: \"kubernetes.io/projected/86078d92-4584-4ed5-9d3b-280e6678e18d-kube-api-access-wvrkd\") pod \"86078d92-4584-4ed5-9d3b-280e6678e18d\" (UID: \"86078d92-4584-4ed5-9d3b-280e6678e18d\") " Oct 13 09:00:46 crc kubenswrapper[4685]: I1013 09:00:46.293016 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86078d92-4584-4ed5-9d3b-280e6678e18d-kube-api-access-wvrkd" (OuterVolumeSpecName: "kube-api-access-wvrkd") pod "86078d92-4584-4ed5-9d3b-280e6678e18d" (UID: "86078d92-4584-4ed5-9d3b-280e6678e18d"). InnerVolumeSpecName "kube-api-access-wvrkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:00:46 crc kubenswrapper[4685]: I1013 09:00:46.388207 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvrkd\" (UniqueName: \"kubernetes.io/projected/86078d92-4584-4ed5-9d3b-280e6678e18d-kube-api-access-wvrkd\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:46 crc kubenswrapper[4685]: I1013 09:00:46.896881 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-r4l79" event={"ID":"86078d92-4584-4ed5-9d3b-280e6678e18d","Type":"ContainerDied","Data":"f783cf1eee0c13170cfd5e5113807b7df9cfbf4b70b436dae18476c2a07657af"} Oct 13 09:00:46 crc kubenswrapper[4685]: I1013 09:00:46.898164 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f783cf1eee0c13170cfd5e5113807b7df9cfbf4b70b436dae18476c2a07657af" Oct 13 09:00:46 crc kubenswrapper[4685]: I1013 09:00:46.897005 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-r4l79" Oct 13 09:00:47 crc kubenswrapper[4685]: I1013 09:00:47.530382 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-990a-account-create-8gxck"] Oct 13 09:00:47 crc kubenswrapper[4685]: E1013 09:00:47.531015 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86078d92-4584-4ed5-9d3b-280e6678e18d" containerName="mariadb-database-create" Oct 13 09:00:47 crc kubenswrapper[4685]: I1013 09:00:47.531046 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="86078d92-4584-4ed5-9d3b-280e6678e18d" containerName="mariadb-database-create" Oct 13 09:00:47 crc kubenswrapper[4685]: I1013 09:00:47.531794 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="86078d92-4584-4ed5-9d3b-280e6678e18d" containerName="mariadb-database-create" Oct 13 09:00:47 crc kubenswrapper[4685]: I1013 09:00:47.532874 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-990a-account-create-8gxck"] Oct 13 09:00:47 crc kubenswrapper[4685]: I1013 09:00:47.533071 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-990a-account-create-8gxck" Oct 13 09:00:47 crc kubenswrapper[4685]: I1013 09:00:47.536513 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 13 09:00:47 crc kubenswrapper[4685]: I1013 09:00:47.613447 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjl72\" (UniqueName: \"kubernetes.io/projected/67ada508-f859-47b2-b17d-009d71fd1af7-kube-api-access-xjl72\") pod \"keystone-990a-account-create-8gxck\" (UID: \"67ada508-f859-47b2-b17d-009d71fd1af7\") " pod="openstack/keystone-990a-account-create-8gxck" Oct 13 09:00:47 crc kubenswrapper[4685]: I1013 09:00:47.715369 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjl72\" (UniqueName: \"kubernetes.io/projected/67ada508-f859-47b2-b17d-009d71fd1af7-kube-api-access-xjl72\") pod \"keystone-990a-account-create-8gxck\" (UID: \"67ada508-f859-47b2-b17d-009d71fd1af7\") " pod="openstack/keystone-990a-account-create-8gxck" Oct 13 09:00:47 crc kubenswrapper[4685]: I1013 09:00:47.742570 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjl72\" (UniqueName: \"kubernetes.io/projected/67ada508-f859-47b2-b17d-009d71fd1af7-kube-api-access-xjl72\") pod \"keystone-990a-account-create-8gxck\" (UID: \"67ada508-f859-47b2-b17d-009d71fd1af7\") " pod="openstack/keystone-990a-account-create-8gxck" Oct 13 09:00:47 crc kubenswrapper[4685]: I1013 09:00:47.857743 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-990a-account-create-8gxck" Oct 13 09:00:47 crc kubenswrapper[4685]: I1013 09:00:47.876217 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-37cf-account-create-n4sgt"] Oct 13 09:00:47 crc kubenswrapper[4685]: I1013 09:00:47.877511 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-37cf-account-create-n4sgt" Oct 13 09:00:47 crc kubenswrapper[4685]: I1013 09:00:47.881859 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-37cf-account-create-n4sgt"] Oct 13 09:00:47 crc kubenswrapper[4685]: I1013 09:00:47.882198 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 13 09:00:47 crc kubenswrapper[4685]: I1013 09:00:47.919528 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lszw\" (UniqueName: \"kubernetes.io/projected/ade16f9d-2076-4f98-ac8b-96e919660fb0-kube-api-access-5lszw\") pod \"placement-37cf-account-create-n4sgt\" (UID: \"ade16f9d-2076-4f98-ac8b-96e919660fb0\") " pod="openstack/placement-37cf-account-create-n4sgt" Oct 13 09:00:48 crc kubenswrapper[4685]: I1013 09:00:48.021203 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lszw\" (UniqueName: \"kubernetes.io/projected/ade16f9d-2076-4f98-ac8b-96e919660fb0-kube-api-access-5lszw\") pod \"placement-37cf-account-create-n4sgt\" (UID: \"ade16f9d-2076-4f98-ac8b-96e919660fb0\") " pod="openstack/placement-37cf-account-create-n4sgt" Oct 13 09:00:48 crc kubenswrapper[4685]: I1013 09:00:48.039559 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lszw\" (UniqueName: \"kubernetes.io/projected/ade16f9d-2076-4f98-ac8b-96e919660fb0-kube-api-access-5lszw\") pod \"placement-37cf-account-create-n4sgt\" (UID: \"ade16f9d-2076-4f98-ac8b-96e919660fb0\") " pod="openstack/placement-37cf-account-create-n4sgt" Oct 13 09:00:48 crc kubenswrapper[4685]: I1013 09:00:48.264162 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-37cf-account-create-n4sgt" Oct 13 09:00:48 crc kubenswrapper[4685]: I1013 09:00:48.352410 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-990a-account-create-8gxck"] Oct 13 09:00:48 crc kubenswrapper[4685]: I1013 09:00:48.835294 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-etc-swift\") pod \"swift-storage-0\" (UID: \"0791156b-11f2-43ad-b910-3e42b4d6670f\") " pod="openstack/swift-storage-0" Oct 13 09:00:48 crc kubenswrapper[4685]: E1013 09:00:48.835476 4685 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 13 09:00:48 crc kubenswrapper[4685]: E1013 09:00:48.835649 4685 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 13 09:00:48 crc kubenswrapper[4685]: E1013 09:00:48.835707 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-etc-swift podName:0791156b-11f2-43ad-b910-3e42b4d6670f nodeName:}" failed. No retries permitted until 2025-10-13 09:00:56.835690733 +0000 UTC m=+981.983566494 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-etc-swift") pod "swift-storage-0" (UID: "0791156b-11f2-43ad-b910-3e42b4d6670f") : configmap "swift-ring-files" not found Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.229587 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.296757 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-szmsc"] Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.297351 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" podUID="9ce22efc-b48d-4d42-a01a-31f71c53c3ab" containerName="dnsmasq-dns" containerID="cri-o://84ce7f67e60c874cfdcb1883fd997c33a283a503994c815b20be8f0c1ae4a60c" gracePeriod=10 Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.635787 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-37cf-account-create-n4sgt"] Oct 13 09:00:50 crc kubenswrapper[4685]: W1013 09:00:50.637019 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podade16f9d_2076_4f98_ac8b_96e919660fb0.slice/crio-bbf9046fa4ef8bfea76401d57f20b2bc95b5ccb9817b0fd2b5c2459a316483a3 WatchSource:0}: Error finding container bbf9046fa4ef8bfea76401d57f20b2bc95b5ccb9817b0fd2b5c2459a316483a3: Status 404 returned error can't find the container with id bbf9046fa4ef8bfea76401d57f20b2bc95b5ccb9817b0fd2b5c2459a316483a3 Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.759219 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.867206 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-dns-svc\") pod \"9ce22efc-b48d-4d42-a01a-31f71c53c3ab\" (UID: \"9ce22efc-b48d-4d42-a01a-31f71c53c3ab\") " Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.867269 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q56cd\" (UniqueName: \"kubernetes.io/projected/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-kube-api-access-q56cd\") pod \"9ce22efc-b48d-4d42-a01a-31f71c53c3ab\" (UID: \"9ce22efc-b48d-4d42-a01a-31f71c53c3ab\") " Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.867390 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-config\") pod \"9ce22efc-b48d-4d42-a01a-31f71c53c3ab\" (UID: \"9ce22efc-b48d-4d42-a01a-31f71c53c3ab\") " Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.867422 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-ovsdbserver-nb\") pod \"9ce22efc-b48d-4d42-a01a-31f71c53c3ab\" (UID: \"9ce22efc-b48d-4d42-a01a-31f71c53c3ab\") " Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.877499 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-kube-api-access-q56cd" (OuterVolumeSpecName: "kube-api-access-q56cd") pod "9ce22efc-b48d-4d42-a01a-31f71c53c3ab" (UID: "9ce22efc-b48d-4d42-a01a-31f71c53c3ab"). InnerVolumeSpecName "kube-api-access-q56cd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.928536 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-config" (OuterVolumeSpecName: "config") pod "9ce22efc-b48d-4d42-a01a-31f71c53c3ab" (UID: "9ce22efc-b48d-4d42-a01a-31f71c53c3ab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.939977 4685 generic.go:334] "Generic (PLEG): container finished" podID="9ce22efc-b48d-4d42-a01a-31f71c53c3ab" containerID="84ce7f67e60c874cfdcb1883fd997c33a283a503994c815b20be8f0c1ae4a60c" exitCode=0 Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.940183 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.940167 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" event={"ID":"9ce22efc-b48d-4d42-a01a-31f71c53c3ab","Type":"ContainerDied","Data":"84ce7f67e60c874cfdcb1883fd997c33a283a503994c815b20be8f0c1ae4a60c"} Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.941038 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-szmsc" event={"ID":"9ce22efc-b48d-4d42-a01a-31f71c53c3ab","Type":"ContainerDied","Data":"0357011fddf91c251e92980dbcfac2fa4f7c87faa594714b40ad5587e7e75edc"} Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.941113 4685 scope.go:117] "RemoveContainer" containerID="84ce7f67e60c874cfdcb1883fd997c33a283a503994c815b20be8f0c1ae4a60c" Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.942764 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-2qrrt" event={"ID":"b884b6de-f048-4f14-b8b0-4775fa1d4cd1","Type":"ContainerStarted","Data":"4a3392cbebf9ac615bab4e3433249580812babd0f3b3b841e3d091fa73fc3710"} Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.943066 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9ce22efc-b48d-4d42-a01a-31f71c53c3ab" (UID: "9ce22efc-b48d-4d42-a01a-31f71c53c3ab"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.946142 4685 generic.go:334] "Generic (PLEG): container finished" podID="3b0348a6-6814-47f3-919e-e975c01e00e8" containerID="a8c575675b8a8240f9a9cce662be0dbf91b784093b2ec63f6e3d873f5d79efc2" exitCode=0 Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.946260 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3b0348a6-6814-47f3-919e-e975c01e00e8","Type":"ContainerDied","Data":"a8c575675b8a8240f9a9cce662be0dbf91b784093b2ec63f6e3d873f5d79efc2"} Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.949951 4685 generic.go:334] "Generic (PLEG): container finished" podID="b7f19239-a2d9-49b5-80b0-43543daf1787" containerID="a6d714e0fc56e7256e93150fb3ec702f4835022733af8f3f1d12eecd1ec2d20b" exitCode=0 Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.950048 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b7f19239-a2d9-49b5-80b0-43543daf1787","Type":"ContainerDied","Data":"a6d714e0fc56e7256e93150fb3ec702f4835022733af8f3f1d12eecd1ec2d20b"} Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.956060 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9ce22efc-b48d-4d42-a01a-31f71c53c3ab" (UID: "9ce22efc-b48d-4d42-a01a-31f71c53c3ab"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.956926 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-37cf-account-create-n4sgt" event={"ID":"ade16f9d-2076-4f98-ac8b-96e919660fb0","Type":"ContainerStarted","Data":"bf9f33929009619aa40e41abc4290323ad182d516af16469850e9a0ebbefd71b"} Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.957026 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-37cf-account-create-n4sgt" event={"ID":"ade16f9d-2076-4f98-ac8b-96e919660fb0","Type":"ContainerStarted","Data":"bbf9046fa4ef8bfea76401d57f20b2bc95b5ccb9817b0fd2b5c2459a316483a3"} Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.966330 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-2qrrt" podStartSLOduration=2.42130275 podStartE2EDuration="6.966310705s" podCreationTimestamp="2025-10-13 09:00:44 +0000 UTC" firstStartedPulling="2025-10-13 09:00:45.687574683 +0000 UTC m=+970.835450444" lastFinishedPulling="2025-10-13 09:00:50.232582638 +0000 UTC m=+975.380458399" observedRunningTime="2025-10-13 09:00:50.962362178 +0000 UTC m=+976.110237939" watchObservedRunningTime="2025-10-13 09:00:50.966310705 +0000 UTC m=+976.114186466" Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.968852 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.968885 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q56cd\" (UniqueName: \"kubernetes.io/projected/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-kube-api-access-q56cd\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.968897 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-config\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.968905 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ce22efc-b48d-4d42-a01a-31f71c53c3ab-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.970689 4685 generic.go:334] "Generic (PLEG): container finished" podID="67ada508-f859-47b2-b17d-009d71fd1af7" containerID="34486ea1191b0d949f55df7e09a19546caebcbcc91c252014519208d976df5eb" exitCode=0 Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.970727 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-990a-account-create-8gxck" event={"ID":"67ada508-f859-47b2-b17d-009d71fd1af7","Type":"ContainerDied","Data":"34486ea1191b0d949f55df7e09a19546caebcbcc91c252014519208d976df5eb"} Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.970748 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-990a-account-create-8gxck" event={"ID":"67ada508-f859-47b2-b17d-009d71fd1af7","Type":"ContainerStarted","Data":"9289101d39421fd393a8c7f3f8851158e6f457d315399b6f6f90dd54fa3f97af"} Oct 13 09:00:50 crc kubenswrapper[4685]: I1013 09:00:50.999111 4685 scope.go:117] "RemoveContainer" containerID="63fc65662155bc9c74d683ff726f68ca952d2e0c78e81c7b946bcc50dba4da94" Oct 13 09:00:51 crc kubenswrapper[4685]: I1013 09:00:51.028043 4685 scope.go:117] "RemoveContainer" containerID="84ce7f67e60c874cfdcb1883fd997c33a283a503994c815b20be8f0c1ae4a60c" Oct 13 09:00:51 crc kubenswrapper[4685]: E1013 09:00:51.028596 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84ce7f67e60c874cfdcb1883fd997c33a283a503994c815b20be8f0c1ae4a60c\": container with ID starting with 84ce7f67e60c874cfdcb1883fd997c33a283a503994c815b20be8f0c1ae4a60c not found: ID does not exist" containerID="84ce7f67e60c874cfdcb1883fd997c33a283a503994c815b20be8f0c1ae4a60c" Oct 13 09:00:51 crc kubenswrapper[4685]: I1013 09:00:51.028650 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84ce7f67e60c874cfdcb1883fd997c33a283a503994c815b20be8f0c1ae4a60c"} err="failed to get container status \"84ce7f67e60c874cfdcb1883fd997c33a283a503994c815b20be8f0c1ae4a60c\": rpc error: code = NotFound desc = could not find container \"84ce7f67e60c874cfdcb1883fd997c33a283a503994c815b20be8f0c1ae4a60c\": container with ID starting with 84ce7f67e60c874cfdcb1883fd997c33a283a503994c815b20be8f0c1ae4a60c not found: ID does not exist" Oct 13 09:00:51 crc kubenswrapper[4685]: I1013 09:00:51.028681 4685 scope.go:117] "RemoveContainer" containerID="63fc65662155bc9c74d683ff726f68ca952d2e0c78e81c7b946bcc50dba4da94" Oct 13 09:00:51 crc kubenswrapper[4685]: E1013 09:00:51.047124 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63fc65662155bc9c74d683ff726f68ca952d2e0c78e81c7b946bcc50dba4da94\": container with ID starting with 63fc65662155bc9c74d683ff726f68ca952d2e0c78e81c7b946bcc50dba4da94 not found: ID does not exist" containerID="63fc65662155bc9c74d683ff726f68ca952d2e0c78e81c7b946bcc50dba4da94" Oct 13 09:00:51 crc kubenswrapper[4685]: I1013 09:00:51.052301 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63fc65662155bc9c74d683ff726f68ca952d2e0c78e81c7b946bcc50dba4da94"} err="failed to get container status \"63fc65662155bc9c74d683ff726f68ca952d2e0c78e81c7b946bcc50dba4da94\": rpc error: code = NotFound desc = could not find container \"63fc65662155bc9c74d683ff726f68ca952d2e0c78e81c7b946bcc50dba4da94\": container with ID starting with 63fc65662155bc9c74d683ff726f68ca952d2e0c78e81c7b946bcc50dba4da94 not found: ID does not exist" Oct 13 09:00:51 crc kubenswrapper[4685]: I1013 09:00:51.272142 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-szmsc"] Oct 13 09:00:51 crc kubenswrapper[4685]: I1013 09:00:51.280418 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-szmsc"] Oct 13 09:00:51 crc kubenswrapper[4685]: I1013 09:00:51.513300 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ce22efc-b48d-4d42-a01a-31f71c53c3ab" path="/var/lib/kubelet/pods/9ce22efc-b48d-4d42-a01a-31f71c53c3ab/volumes" Oct 13 09:00:51 crc kubenswrapper[4685]: I1013 09:00:51.978248 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b7f19239-a2d9-49b5-80b0-43543daf1787","Type":"ContainerStarted","Data":"0ab5c1c1eaab4696cc58a8975e88c605263b37c6525d8d978b6c52d12ae3182e"} Oct 13 09:00:51 crc kubenswrapper[4685]: I1013 09:00:51.978465 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:00:51 crc kubenswrapper[4685]: I1013 09:00:51.979738 4685 generic.go:334] "Generic (PLEG): container finished" podID="ade16f9d-2076-4f98-ac8b-96e919660fb0" containerID="bf9f33929009619aa40e41abc4290323ad182d516af16469850e9a0ebbefd71b" exitCode=0 Oct 13 09:00:51 crc kubenswrapper[4685]: I1013 09:00:51.979766 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-37cf-account-create-n4sgt" event={"ID":"ade16f9d-2076-4f98-ac8b-96e919660fb0","Type":"ContainerDied","Data":"bf9f33929009619aa40e41abc4290323ad182d516af16469850e9a0ebbefd71b"} Oct 13 09:00:51 crc kubenswrapper[4685]: I1013 09:00:51.983478 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3b0348a6-6814-47f3-919e-e975c01e00e8","Type":"ContainerStarted","Data":"692441bc08af17df5ba27504a2d82f321da33197cff11964acc63b131225301e"} Oct 13 09:00:51 crc kubenswrapper[4685]: I1013 09:00:51.984205 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 13 09:00:52 crc kubenswrapper[4685]: I1013 09:00:52.004258 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.209699243 podStartE2EDuration="59.004242197s" podCreationTimestamp="2025-10-13 08:59:53 +0000 UTC" firstStartedPulling="2025-10-13 08:59:55.685977156 +0000 UTC m=+920.833852917" lastFinishedPulling="2025-10-13 09:00:17.48052009 +0000 UTC m=+942.628395871" observedRunningTime="2025-10-13 09:00:52.003433395 +0000 UTC m=+977.151309156" watchObservedRunningTime="2025-10-13 09:00:52.004242197 +0000 UTC m=+977.152117958" Oct 13 09:00:52 crc kubenswrapper[4685]: I1013 09:00:52.035733 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=-9223371977.819065 podStartE2EDuration="59.035710336s" podCreationTimestamp="2025-10-13 08:59:53 +0000 UTC" firstStartedPulling="2025-10-13 08:59:55.658662318 +0000 UTC m=+920.806538079" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:00:52.029114436 +0000 UTC m=+977.176990207" watchObservedRunningTime="2025-10-13 09:00:52.035710336 +0000 UTC m=+977.183586097" Oct 13 09:00:52 crc kubenswrapper[4685]: I1013 09:00:52.427778 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-37cf-account-create-n4sgt" Oct 13 09:00:52 crc kubenswrapper[4685]: I1013 09:00:52.437471 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-990a-account-create-8gxck" Oct 13 09:00:52 crc kubenswrapper[4685]: I1013 09:00:52.492820 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjl72\" (UniqueName: \"kubernetes.io/projected/67ada508-f859-47b2-b17d-009d71fd1af7-kube-api-access-xjl72\") pod \"67ada508-f859-47b2-b17d-009d71fd1af7\" (UID: \"67ada508-f859-47b2-b17d-009d71fd1af7\") " Oct 13 09:00:52 crc kubenswrapper[4685]: I1013 09:00:52.493013 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lszw\" (UniqueName: \"kubernetes.io/projected/ade16f9d-2076-4f98-ac8b-96e919660fb0-kube-api-access-5lszw\") pod \"ade16f9d-2076-4f98-ac8b-96e919660fb0\" (UID: \"ade16f9d-2076-4f98-ac8b-96e919660fb0\") " Oct 13 09:00:52 crc kubenswrapper[4685]: I1013 09:00:52.499306 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ade16f9d-2076-4f98-ac8b-96e919660fb0-kube-api-access-5lszw" (OuterVolumeSpecName: "kube-api-access-5lszw") pod "ade16f9d-2076-4f98-ac8b-96e919660fb0" (UID: "ade16f9d-2076-4f98-ac8b-96e919660fb0"). InnerVolumeSpecName "kube-api-access-5lszw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:00:52 crc kubenswrapper[4685]: I1013 09:00:52.504352 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67ada508-f859-47b2-b17d-009d71fd1af7-kube-api-access-xjl72" (OuterVolumeSpecName: "kube-api-access-xjl72") pod "67ada508-f859-47b2-b17d-009d71fd1af7" (UID: "67ada508-f859-47b2-b17d-009d71fd1af7"). InnerVolumeSpecName "kube-api-access-xjl72". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:00:52 crc kubenswrapper[4685]: I1013 09:00:52.594334 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lszw\" (UniqueName: \"kubernetes.io/projected/ade16f9d-2076-4f98-ac8b-96e919660fb0-kube-api-access-5lszw\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:52 crc kubenswrapper[4685]: I1013 09:00:52.594365 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjl72\" (UniqueName: \"kubernetes.io/projected/67ada508-f859-47b2-b17d-009d71fd1af7-kube-api-access-xjl72\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:52 crc kubenswrapper[4685]: I1013 09:00:52.601767 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 13 09:00:52 crc kubenswrapper[4685]: I1013 09:00:52.991812 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-990a-account-create-8gxck" event={"ID":"67ada508-f859-47b2-b17d-009d71fd1af7","Type":"ContainerDied","Data":"9289101d39421fd393a8c7f3f8851158e6f457d315399b6f6f90dd54fa3f97af"} Oct 13 09:00:52 crc kubenswrapper[4685]: I1013 09:00:52.991852 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-990a-account-create-8gxck" Oct 13 09:00:52 crc kubenswrapper[4685]: I1013 09:00:52.991858 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9289101d39421fd393a8c7f3f8851158e6f457d315399b6f6f90dd54fa3f97af" Oct 13 09:00:52 crc kubenswrapper[4685]: I1013 09:00:52.993710 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-37cf-account-create-n4sgt" event={"ID":"ade16f9d-2076-4f98-ac8b-96e919660fb0","Type":"ContainerDied","Data":"bbf9046fa4ef8bfea76401d57f20b2bc95b5ccb9817b0fd2b5c2459a316483a3"} Oct 13 09:00:52 crc kubenswrapper[4685]: I1013 09:00:52.993738 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbf9046fa4ef8bfea76401d57f20b2bc95b5ccb9817b0fd2b5c2459a316483a3" Oct 13 09:00:52 crc kubenswrapper[4685]: I1013 09:00:52.993803 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-37cf-account-create-n4sgt" Oct 13 09:00:53 crc kubenswrapper[4685]: I1013 09:00:53.326983 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-699e-account-create-8b2g4"] Oct 13 09:00:53 crc kubenswrapper[4685]: E1013 09:00:53.327278 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67ada508-f859-47b2-b17d-009d71fd1af7" containerName="mariadb-account-create" Oct 13 09:00:53 crc kubenswrapper[4685]: I1013 09:00:53.327294 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="67ada508-f859-47b2-b17d-009d71fd1af7" containerName="mariadb-account-create" Oct 13 09:00:53 crc kubenswrapper[4685]: E1013 09:00:53.327304 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ade16f9d-2076-4f98-ac8b-96e919660fb0" containerName="mariadb-account-create" Oct 13 09:00:53 crc kubenswrapper[4685]: I1013 09:00:53.327311 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ade16f9d-2076-4f98-ac8b-96e919660fb0" containerName="mariadb-account-create" Oct 13 09:00:53 crc kubenswrapper[4685]: E1013 09:00:53.327340 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ce22efc-b48d-4d42-a01a-31f71c53c3ab" containerName="init" Oct 13 09:00:53 crc kubenswrapper[4685]: I1013 09:00:53.327346 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ce22efc-b48d-4d42-a01a-31f71c53c3ab" containerName="init" Oct 13 09:00:53 crc kubenswrapper[4685]: E1013 09:00:53.327364 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ce22efc-b48d-4d42-a01a-31f71c53c3ab" containerName="dnsmasq-dns" Oct 13 09:00:53 crc kubenswrapper[4685]: I1013 09:00:53.327370 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ce22efc-b48d-4d42-a01a-31f71c53c3ab" containerName="dnsmasq-dns" Oct 13 09:00:53 crc kubenswrapper[4685]: I1013 09:00:53.327518 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ade16f9d-2076-4f98-ac8b-96e919660fb0" containerName="mariadb-account-create" Oct 13 09:00:53 crc kubenswrapper[4685]: I1013 09:00:53.327532 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="67ada508-f859-47b2-b17d-009d71fd1af7" containerName="mariadb-account-create" Oct 13 09:00:53 crc kubenswrapper[4685]: I1013 09:00:53.327541 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ce22efc-b48d-4d42-a01a-31f71c53c3ab" containerName="dnsmasq-dns" Oct 13 09:00:53 crc kubenswrapper[4685]: I1013 09:00:53.328037 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-699e-account-create-8b2g4" Oct 13 09:00:53 crc kubenswrapper[4685]: I1013 09:00:53.330805 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 13 09:00:53 crc kubenswrapper[4685]: I1013 09:00:53.345768 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-699e-account-create-8b2g4"] Oct 13 09:00:53 crc kubenswrapper[4685]: I1013 09:00:53.415546 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sl2f2\" (UniqueName: \"kubernetes.io/projected/24a0a6b1-caf0-4256-b80d-0c6d8171f2de-kube-api-access-sl2f2\") pod \"glance-699e-account-create-8b2g4\" (UID: \"24a0a6b1-caf0-4256-b80d-0c6d8171f2de\") " pod="openstack/glance-699e-account-create-8b2g4" Oct 13 09:00:53 crc kubenswrapper[4685]: I1013 09:00:53.516881 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sl2f2\" (UniqueName: \"kubernetes.io/projected/24a0a6b1-caf0-4256-b80d-0c6d8171f2de-kube-api-access-sl2f2\") pod \"glance-699e-account-create-8b2g4\" (UID: \"24a0a6b1-caf0-4256-b80d-0c6d8171f2de\") " pod="openstack/glance-699e-account-create-8b2g4" Oct 13 09:00:53 crc kubenswrapper[4685]: I1013 09:00:53.537654 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sl2f2\" (UniqueName: \"kubernetes.io/projected/24a0a6b1-caf0-4256-b80d-0c6d8171f2de-kube-api-access-sl2f2\") pod \"glance-699e-account-create-8b2g4\" (UID: \"24a0a6b1-caf0-4256-b80d-0c6d8171f2de\") " pod="openstack/glance-699e-account-create-8b2g4" Oct 13 09:00:53 crc kubenswrapper[4685]: I1013 09:00:53.649628 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-699e-account-create-8b2g4" Oct 13 09:00:53 crc kubenswrapper[4685]: I1013 09:00:53.889160 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-699e-account-create-8b2g4"] Oct 13 09:00:54 crc kubenswrapper[4685]: I1013 09:00:54.002239 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-699e-account-create-8b2g4" event={"ID":"24a0a6b1-caf0-4256-b80d-0c6d8171f2de","Type":"ContainerStarted","Data":"21c7bd3c389ecf37d97596049a457a3bb12d448edc3e9d4175e535068ad48d48"} Oct 13 09:00:55 crc kubenswrapper[4685]: I1013 09:00:55.008635 4685 generic.go:334] "Generic (PLEG): container finished" podID="24a0a6b1-caf0-4256-b80d-0c6d8171f2de" containerID="31e1a7014e1b7c1cec0638f25f223a1b5e50083c2510c2962394510cd25341cb" exitCode=0 Oct 13 09:00:55 crc kubenswrapper[4685]: I1013 09:00:55.008673 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-699e-account-create-8b2g4" event={"ID":"24a0a6b1-caf0-4256-b80d-0c6d8171f2de","Type":"ContainerDied","Data":"31e1a7014e1b7c1cec0638f25f223a1b5e50083c2510c2962394510cd25341cb"} Oct 13 09:00:56 crc kubenswrapper[4685]: I1013 09:00:56.385918 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-699e-account-create-8b2g4" Oct 13 09:00:56 crc kubenswrapper[4685]: I1013 09:00:56.579395 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sl2f2\" (UniqueName: \"kubernetes.io/projected/24a0a6b1-caf0-4256-b80d-0c6d8171f2de-kube-api-access-sl2f2\") pod \"24a0a6b1-caf0-4256-b80d-0c6d8171f2de\" (UID: \"24a0a6b1-caf0-4256-b80d-0c6d8171f2de\") " Oct 13 09:00:56 crc kubenswrapper[4685]: I1013 09:00:56.588342 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24a0a6b1-caf0-4256-b80d-0c6d8171f2de-kube-api-access-sl2f2" (OuterVolumeSpecName: "kube-api-access-sl2f2") pod "24a0a6b1-caf0-4256-b80d-0c6d8171f2de" (UID: "24a0a6b1-caf0-4256-b80d-0c6d8171f2de"). InnerVolumeSpecName "kube-api-access-sl2f2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:00:56 crc kubenswrapper[4685]: I1013 09:00:56.682031 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sl2f2\" (UniqueName: \"kubernetes.io/projected/24a0a6b1-caf0-4256-b80d-0c6d8171f2de-kube-api-access-sl2f2\") on node \"crc\" DevicePath \"\"" Oct 13 09:00:56 crc kubenswrapper[4685]: E1013 09:00:56.885872 4685 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 13 09:00:56 crc kubenswrapper[4685]: E1013 09:00:56.885906 4685 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 13 09:00:56 crc kubenswrapper[4685]: E1013 09:00:56.885986 4685 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-etc-swift podName:0791156b-11f2-43ad-b910-3e42b4d6670f nodeName:}" failed. No retries permitted until 2025-10-13 09:01:12.885965409 +0000 UTC m=+998.033841170 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-etc-swift") pod "swift-storage-0" (UID: "0791156b-11f2-43ad-b910-3e42b4d6670f") : configmap "swift-ring-files" not found Oct 13 09:00:56 crc kubenswrapper[4685]: I1013 09:00:56.886325 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-etc-swift\") pod \"swift-storage-0\" (UID: \"0791156b-11f2-43ad-b910-3e42b4d6670f\") " pod="openstack/swift-storage-0" Oct 13 09:00:57 crc kubenswrapper[4685]: I1013 09:00:57.022668 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-699e-account-create-8b2g4" event={"ID":"24a0a6b1-caf0-4256-b80d-0c6d8171f2de","Type":"ContainerDied","Data":"21c7bd3c389ecf37d97596049a457a3bb12d448edc3e9d4175e535068ad48d48"} Oct 13 09:00:57 crc kubenswrapper[4685]: I1013 09:00:57.022903 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21c7bd3c389ecf37d97596049a457a3bb12d448edc3e9d4175e535068ad48d48" Oct 13 09:00:57 crc kubenswrapper[4685]: I1013 09:00:57.022963 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-699e-account-create-8b2g4" Oct 13 09:00:58 crc kubenswrapper[4685]: I1013 09:00:58.478852 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-jvxxs"] Oct 13 09:00:58 crc kubenswrapper[4685]: E1013 09:00:58.479287 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24a0a6b1-caf0-4256-b80d-0c6d8171f2de" containerName="mariadb-account-create" Oct 13 09:00:58 crc kubenswrapper[4685]: I1013 09:00:58.479303 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="24a0a6b1-caf0-4256-b80d-0c6d8171f2de" containerName="mariadb-account-create" Oct 13 09:00:58 crc kubenswrapper[4685]: I1013 09:00:58.479444 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="24a0a6b1-caf0-4256-b80d-0c6d8171f2de" containerName="mariadb-account-create" Oct 13 09:00:58 crc kubenswrapper[4685]: I1013 09:00:58.480038 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jvxxs" Oct 13 09:00:58 crc kubenswrapper[4685]: I1013 09:00:58.486605 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gt2pf" Oct 13 09:00:58 crc kubenswrapper[4685]: I1013 09:00:58.491677 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 13 09:00:58 crc kubenswrapper[4685]: I1013 09:00:58.496006 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-jvxxs"] Oct 13 09:00:58 crc kubenswrapper[4685]: I1013 09:00:58.528376 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-config-data\") pod \"glance-db-sync-jvxxs\" (UID: \"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251\") " pod="openstack/glance-db-sync-jvxxs" Oct 13 09:00:58 crc kubenswrapper[4685]: I1013 09:00:58.528438 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-db-sync-config-data\") pod \"glance-db-sync-jvxxs\" (UID: \"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251\") " pod="openstack/glance-db-sync-jvxxs" Oct 13 09:00:58 crc kubenswrapper[4685]: I1013 09:00:58.528556 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cthk7\" (UniqueName: \"kubernetes.io/projected/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-kube-api-access-cthk7\") pod \"glance-db-sync-jvxxs\" (UID: \"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251\") " pod="openstack/glance-db-sync-jvxxs" Oct 13 09:00:58 crc kubenswrapper[4685]: I1013 09:00:58.528617 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-combined-ca-bundle\") pod \"glance-db-sync-jvxxs\" (UID: \"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251\") " pod="openstack/glance-db-sync-jvxxs" Oct 13 09:00:58 crc kubenswrapper[4685]: I1013 09:00:58.629747 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-combined-ca-bundle\") pod \"glance-db-sync-jvxxs\" (UID: \"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251\") " pod="openstack/glance-db-sync-jvxxs" Oct 13 09:00:58 crc kubenswrapper[4685]: I1013 09:00:58.629909 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-config-data\") pod \"glance-db-sync-jvxxs\" (UID: \"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251\") " pod="openstack/glance-db-sync-jvxxs" Oct 13 09:00:58 crc kubenswrapper[4685]: I1013 09:00:58.629964 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-db-sync-config-data\") pod \"glance-db-sync-jvxxs\" (UID: \"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251\") " pod="openstack/glance-db-sync-jvxxs" Oct 13 09:00:58 crc kubenswrapper[4685]: I1013 09:00:58.630021 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cthk7\" (UniqueName: \"kubernetes.io/projected/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-kube-api-access-cthk7\") pod \"glance-db-sync-jvxxs\" (UID: \"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251\") " pod="openstack/glance-db-sync-jvxxs" Oct 13 09:00:58 crc kubenswrapper[4685]: I1013 09:00:58.637698 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-db-sync-config-data\") pod \"glance-db-sync-jvxxs\" (UID: \"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251\") " pod="openstack/glance-db-sync-jvxxs" Oct 13 09:00:58 crc kubenswrapper[4685]: I1013 09:00:58.637752 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-combined-ca-bundle\") pod \"glance-db-sync-jvxxs\" (UID: \"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251\") " pod="openstack/glance-db-sync-jvxxs" Oct 13 09:00:58 crc kubenswrapper[4685]: I1013 09:00:58.647942 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-config-data\") pod \"glance-db-sync-jvxxs\" (UID: \"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251\") " pod="openstack/glance-db-sync-jvxxs" Oct 13 09:00:58 crc kubenswrapper[4685]: I1013 09:00:58.668349 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cthk7\" (UniqueName: \"kubernetes.io/projected/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-kube-api-access-cthk7\") pod \"glance-db-sync-jvxxs\" (UID: \"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251\") " pod="openstack/glance-db-sync-jvxxs" Oct 13 09:00:58 crc kubenswrapper[4685]: I1013 09:00:58.812041 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jvxxs" Oct 13 09:00:59 crc kubenswrapper[4685]: I1013 09:00:59.043717 4685 generic.go:334] "Generic (PLEG): container finished" podID="b884b6de-f048-4f14-b8b0-4775fa1d4cd1" containerID="4a3392cbebf9ac615bab4e3433249580812babd0f3b3b841e3d091fa73fc3710" exitCode=0 Oct 13 09:00:59 crc kubenswrapper[4685]: I1013 09:00:59.043767 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-2qrrt" event={"ID":"b884b6de-f048-4f14-b8b0-4775fa1d4cd1","Type":"ContainerDied","Data":"4a3392cbebf9ac615bab4e3433249580812babd0f3b3b841e3d091fa73fc3710"} Oct 13 09:00:59 crc kubenswrapper[4685]: W1013 09:00:59.348032 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5a0947d_3515_4e7c_a69f_c7e4d3e9f251.slice/crio-efb67fa5579071bb13f37055a2e4accfee710a1916b236c22b47a52d88b4f77b WatchSource:0}: Error finding container efb67fa5579071bb13f37055a2e4accfee710a1916b236c22b47a52d88b4f77b: Status 404 returned error can't find the container with id efb67fa5579071bb13f37055a2e4accfee710a1916b236c22b47a52d88b4f77b Oct 13 09:00:59 crc kubenswrapper[4685]: I1013 09:00:59.357992 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-jvxxs"] Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.021202 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-tb2c4" podUID="9dccff31-26fc-43c7-a679-01173d2f34a9" containerName="ovn-controller" probeResult="failure" output=< Oct 13 09:01:00 crc kubenswrapper[4685]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 13 09:01:00 crc kubenswrapper[4685]: > Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.038158 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.059750 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jvxxs" event={"ID":"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251","Type":"ContainerStarted","Data":"efb67fa5579071bb13f37055a2e4accfee710a1916b236c22b47a52d88b4f77b"} Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.082398 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-gh7ls" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.321549 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-tb2c4-config-s4tlh"] Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.322753 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tb2c4-config-s4tlh" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.326884 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.339401 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tb2c4-config-s4tlh"] Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.376797 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cd819379-6804-4352-8ad0-416859783cd4-additional-scripts\") pod \"ovn-controller-tb2c4-config-s4tlh\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " pod="openstack/ovn-controller-tb2c4-config-s4tlh" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.376839 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cd819379-6804-4352-8ad0-416859783cd4-var-log-ovn\") pod \"ovn-controller-tb2c4-config-s4tlh\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " pod="openstack/ovn-controller-tb2c4-config-s4tlh" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.376890 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd819379-6804-4352-8ad0-416859783cd4-scripts\") pod \"ovn-controller-tb2c4-config-s4tlh\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " pod="openstack/ovn-controller-tb2c4-config-s4tlh" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.377049 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cd819379-6804-4352-8ad0-416859783cd4-var-run-ovn\") pod \"ovn-controller-tb2c4-config-s4tlh\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " pod="openstack/ovn-controller-tb2c4-config-s4tlh" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.377094 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2plvl\" (UniqueName: \"kubernetes.io/projected/cd819379-6804-4352-8ad0-416859783cd4-kube-api-access-2plvl\") pod \"ovn-controller-tb2c4-config-s4tlh\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " pod="openstack/ovn-controller-tb2c4-config-s4tlh" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.377134 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cd819379-6804-4352-8ad0-416859783cd4-var-run\") pod \"ovn-controller-tb2c4-config-s4tlh\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " pod="openstack/ovn-controller-tb2c4-config-s4tlh" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.410566 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.479738 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-scripts\") pod \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.480073 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5tv7\" (UniqueName: \"kubernetes.io/projected/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-kube-api-access-h5tv7\") pod \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.480194 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-dispersionconf\") pod \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.480215 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-combined-ca-bundle\") pod \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.480271 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-ring-data-devices\") pod \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.480297 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-etc-swift\") pod \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.480316 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-swiftconf\") pod \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\" (UID: \"b884b6de-f048-4f14-b8b0-4775fa1d4cd1\") " Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.480526 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cd819379-6804-4352-8ad0-416859783cd4-var-run\") pod \"ovn-controller-tb2c4-config-s4tlh\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " pod="openstack/ovn-controller-tb2c4-config-s4tlh" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.480565 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cd819379-6804-4352-8ad0-416859783cd4-additional-scripts\") pod \"ovn-controller-tb2c4-config-s4tlh\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " pod="openstack/ovn-controller-tb2c4-config-s4tlh" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.480586 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cd819379-6804-4352-8ad0-416859783cd4-var-log-ovn\") pod \"ovn-controller-tb2c4-config-s4tlh\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " pod="openstack/ovn-controller-tb2c4-config-s4tlh" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.480638 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd819379-6804-4352-8ad0-416859783cd4-scripts\") pod \"ovn-controller-tb2c4-config-s4tlh\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " pod="openstack/ovn-controller-tb2c4-config-s4tlh" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.480738 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cd819379-6804-4352-8ad0-416859783cd4-var-run-ovn\") pod \"ovn-controller-tb2c4-config-s4tlh\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " pod="openstack/ovn-controller-tb2c4-config-s4tlh" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.480789 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2plvl\" (UniqueName: \"kubernetes.io/projected/cd819379-6804-4352-8ad0-416859783cd4-kube-api-access-2plvl\") pod \"ovn-controller-tb2c4-config-s4tlh\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " pod="openstack/ovn-controller-tb2c4-config-s4tlh" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.481095 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cd819379-6804-4352-8ad0-416859783cd4-var-run\") pod \"ovn-controller-tb2c4-config-s4tlh\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " pod="openstack/ovn-controller-tb2c4-config-s4tlh" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.481110 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cd819379-6804-4352-8ad0-416859783cd4-var-log-ovn\") pod \"ovn-controller-tb2c4-config-s4tlh\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " pod="openstack/ovn-controller-tb2c4-config-s4tlh" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.481775 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "b884b6de-f048-4f14-b8b0-4775fa1d4cd1" (UID: "b884b6de-f048-4f14-b8b0-4775fa1d4cd1"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.481820 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cd819379-6804-4352-8ad0-416859783cd4-additional-scripts\") pod \"ovn-controller-tb2c4-config-s4tlh\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " pod="openstack/ovn-controller-tb2c4-config-s4tlh" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.482127 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b884b6de-f048-4f14-b8b0-4775fa1d4cd1" (UID: "b884b6de-f048-4f14-b8b0-4775fa1d4cd1"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.484882 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cd819379-6804-4352-8ad0-416859783cd4-var-run-ovn\") pod \"ovn-controller-tb2c4-config-s4tlh\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " pod="openstack/ovn-controller-tb2c4-config-s4tlh" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.487259 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd819379-6804-4352-8ad0-416859783cd4-scripts\") pod \"ovn-controller-tb2c4-config-s4tlh\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " pod="openstack/ovn-controller-tb2c4-config-s4tlh" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.495129 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-kube-api-access-h5tv7" (OuterVolumeSpecName: "kube-api-access-h5tv7") pod "b884b6de-f048-4f14-b8b0-4775fa1d4cd1" (UID: "b884b6de-f048-4f14-b8b0-4775fa1d4cd1"). InnerVolumeSpecName "kube-api-access-h5tv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.505242 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2plvl\" (UniqueName: \"kubernetes.io/projected/cd819379-6804-4352-8ad0-416859783cd4-kube-api-access-2plvl\") pod \"ovn-controller-tb2c4-config-s4tlh\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " pod="openstack/ovn-controller-tb2c4-config-s4tlh" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.513965 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "b884b6de-f048-4f14-b8b0-4775fa1d4cd1" (UID: "b884b6de-f048-4f14-b8b0-4775fa1d4cd1"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.526281 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-scripts" (OuterVolumeSpecName: "scripts") pod "b884b6de-f048-4f14-b8b0-4775fa1d4cd1" (UID: "b884b6de-f048-4f14-b8b0-4775fa1d4cd1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.559788 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "b884b6de-f048-4f14-b8b0-4775fa1d4cd1" (UID: "b884b6de-f048-4f14-b8b0-4775fa1d4cd1"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.560486 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b884b6de-f048-4f14-b8b0-4775fa1d4cd1" (UID: "b884b6de-f048-4f14-b8b0-4775fa1d4cd1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.583852 4685 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.583886 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.583897 4685 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.583907 4685 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.583934 4685 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.583947 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.583959 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5tv7\" (UniqueName: \"kubernetes.io/projected/b884b6de-f048-4f14-b8b0-4775fa1d4cd1-kube-api-access-h5tv7\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:00 crc kubenswrapper[4685]: I1013 09:01:00.707645 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tb2c4-config-s4tlh" Oct 13 09:01:01 crc kubenswrapper[4685]: I1013 09:01:01.069818 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-2qrrt" event={"ID":"b884b6de-f048-4f14-b8b0-4775fa1d4cd1","Type":"ContainerDied","Data":"ac18b068f1de112c3337f0ccad2c80b24f38a378ed6422cdec7f5e9bf6e711ce"} Oct 13 09:01:01 crc kubenswrapper[4685]: I1013 09:01:01.070086 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac18b068f1de112c3337f0ccad2c80b24f38a378ed6422cdec7f5e9bf6e711ce" Oct 13 09:01:01 crc kubenswrapper[4685]: I1013 09:01:01.069884 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2qrrt" Oct 13 09:01:01 crc kubenswrapper[4685]: I1013 09:01:01.157978 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tb2c4-config-s4tlh"] Oct 13 09:01:02 crc kubenswrapper[4685]: I1013 09:01:02.085829 4685 generic.go:334] "Generic (PLEG): container finished" podID="cd819379-6804-4352-8ad0-416859783cd4" containerID="2abbed46b374258b1c20646f9e96b0579634e4fd1f5cfdf374dbbef34bc7d198" exitCode=0 Oct 13 09:01:02 crc kubenswrapper[4685]: I1013 09:01:02.086005 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tb2c4-config-s4tlh" event={"ID":"cd819379-6804-4352-8ad0-416859783cd4","Type":"ContainerDied","Data":"2abbed46b374258b1c20646f9e96b0579634e4fd1f5cfdf374dbbef34bc7d198"} Oct 13 09:01:02 crc kubenswrapper[4685]: I1013 09:01:02.086139 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tb2c4-config-s4tlh" event={"ID":"cd819379-6804-4352-8ad0-416859783cd4","Type":"ContainerStarted","Data":"bf2691126a69b81cd6ea464c1c7b3765f5f84a0fd59191aec453bee6daae8f22"} Oct 13 09:01:03 crc kubenswrapper[4685]: I1013 09:01:03.381818 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tb2c4-config-s4tlh" Oct 13 09:01:03 crc kubenswrapper[4685]: I1013 09:01:03.458220 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2plvl\" (UniqueName: \"kubernetes.io/projected/cd819379-6804-4352-8ad0-416859783cd4-kube-api-access-2plvl\") pod \"cd819379-6804-4352-8ad0-416859783cd4\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " Oct 13 09:01:03 crc kubenswrapper[4685]: I1013 09:01:03.458290 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd819379-6804-4352-8ad0-416859783cd4-scripts\") pod \"cd819379-6804-4352-8ad0-416859783cd4\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " Oct 13 09:01:03 crc kubenswrapper[4685]: I1013 09:01:03.458315 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cd819379-6804-4352-8ad0-416859783cd4-var-run-ovn\") pod \"cd819379-6804-4352-8ad0-416859783cd4\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " Oct 13 09:01:03 crc kubenswrapper[4685]: I1013 09:01:03.458360 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cd819379-6804-4352-8ad0-416859783cd4-additional-scripts\") pod \"cd819379-6804-4352-8ad0-416859783cd4\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " Oct 13 09:01:03 crc kubenswrapper[4685]: I1013 09:01:03.458440 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cd819379-6804-4352-8ad0-416859783cd4-var-log-ovn\") pod \"cd819379-6804-4352-8ad0-416859783cd4\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " Oct 13 09:01:03 crc kubenswrapper[4685]: I1013 09:01:03.458450 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd819379-6804-4352-8ad0-416859783cd4-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "cd819379-6804-4352-8ad0-416859783cd4" (UID: "cd819379-6804-4352-8ad0-416859783cd4"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 09:01:03 crc kubenswrapper[4685]: I1013 09:01:03.458485 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cd819379-6804-4352-8ad0-416859783cd4-var-run\") pod \"cd819379-6804-4352-8ad0-416859783cd4\" (UID: \"cd819379-6804-4352-8ad0-416859783cd4\") " Oct 13 09:01:03 crc kubenswrapper[4685]: I1013 09:01:03.458543 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd819379-6804-4352-8ad0-416859783cd4-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "cd819379-6804-4352-8ad0-416859783cd4" (UID: "cd819379-6804-4352-8ad0-416859783cd4"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 09:01:03 crc kubenswrapper[4685]: I1013 09:01:03.458714 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd819379-6804-4352-8ad0-416859783cd4-var-run" (OuterVolumeSpecName: "var-run") pod "cd819379-6804-4352-8ad0-416859783cd4" (UID: "cd819379-6804-4352-8ad0-416859783cd4"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 09:01:03 crc kubenswrapper[4685]: I1013 09:01:03.459028 4685 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cd819379-6804-4352-8ad0-416859783cd4-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:03 crc kubenswrapper[4685]: I1013 09:01:03.459051 4685 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cd819379-6804-4352-8ad0-416859783cd4-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:03 crc kubenswrapper[4685]: I1013 09:01:03.459062 4685 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cd819379-6804-4352-8ad0-416859783cd4-var-run\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:03 crc kubenswrapper[4685]: I1013 09:01:03.459209 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd819379-6804-4352-8ad0-416859783cd4-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "cd819379-6804-4352-8ad0-416859783cd4" (UID: "cd819379-6804-4352-8ad0-416859783cd4"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:03 crc kubenswrapper[4685]: I1013 09:01:03.459489 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd819379-6804-4352-8ad0-416859783cd4-scripts" (OuterVolumeSpecName: "scripts") pod "cd819379-6804-4352-8ad0-416859783cd4" (UID: "cd819379-6804-4352-8ad0-416859783cd4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:03 crc kubenswrapper[4685]: I1013 09:01:03.465676 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd819379-6804-4352-8ad0-416859783cd4-kube-api-access-2plvl" (OuterVolumeSpecName: "kube-api-access-2plvl") pod "cd819379-6804-4352-8ad0-416859783cd4" (UID: "cd819379-6804-4352-8ad0-416859783cd4"). InnerVolumeSpecName "kube-api-access-2plvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:01:03 crc kubenswrapper[4685]: I1013 09:01:03.561082 4685 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cd819379-6804-4352-8ad0-416859783cd4-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:03 crc kubenswrapper[4685]: I1013 09:01:03.561360 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2plvl\" (UniqueName: \"kubernetes.io/projected/cd819379-6804-4352-8ad0-416859783cd4-kube-api-access-2plvl\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:03 crc kubenswrapper[4685]: I1013 09:01:03.561373 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd819379-6804-4352-8ad0-416859783cd4-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.108130 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tb2c4-config-s4tlh" event={"ID":"cd819379-6804-4352-8ad0-416859783cd4","Type":"ContainerDied","Data":"bf2691126a69b81cd6ea464c1c7b3765f5f84a0fd59191aec453bee6daae8f22"} Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.108187 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf2691126a69b81cd6ea464c1c7b3765f5f84a0fd59191aec453bee6daae8f22" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.108262 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tb2c4-config-s4tlh" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.489988 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-tb2c4-config-s4tlh"] Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.499264 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-tb2c4-config-s4tlh"] Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.580054 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-tb2c4-config-gb9jj"] Oct 13 09:01:04 crc kubenswrapper[4685]: E1013 09:01:04.580355 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b884b6de-f048-4f14-b8b0-4775fa1d4cd1" containerName="swift-ring-rebalance" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.580371 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b884b6de-f048-4f14-b8b0-4775fa1d4cd1" containerName="swift-ring-rebalance" Oct 13 09:01:04 crc kubenswrapper[4685]: E1013 09:01:04.580394 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd819379-6804-4352-8ad0-416859783cd4" containerName="ovn-config" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.580401 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd819379-6804-4352-8ad0-416859783cd4" containerName="ovn-config" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.580549 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b884b6de-f048-4f14-b8b0-4775fa1d4cd1" containerName="swift-ring-rebalance" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.580567 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd819379-6804-4352-8ad0-416859783cd4" containerName="ovn-config" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.581047 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tb2c4-config-gb9jj" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.583064 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.599149 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tb2c4-config-gb9jj"] Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.680767 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49f6d2fc-6fd2-4823-befc-332a780eb6d4-scripts\") pod \"ovn-controller-tb2c4-config-gb9jj\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " pod="openstack/ovn-controller-tb2c4-config-gb9jj" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.680826 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/49f6d2fc-6fd2-4823-befc-332a780eb6d4-var-run\") pod \"ovn-controller-tb2c4-config-gb9jj\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " pod="openstack/ovn-controller-tb2c4-config-gb9jj" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.680858 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v9zf\" (UniqueName: \"kubernetes.io/projected/49f6d2fc-6fd2-4823-befc-332a780eb6d4-kube-api-access-7v9zf\") pod \"ovn-controller-tb2c4-config-gb9jj\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " pod="openstack/ovn-controller-tb2c4-config-gb9jj" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.680894 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/49f6d2fc-6fd2-4823-befc-332a780eb6d4-var-run-ovn\") pod \"ovn-controller-tb2c4-config-gb9jj\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " pod="openstack/ovn-controller-tb2c4-config-gb9jj" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.680934 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/49f6d2fc-6fd2-4823-befc-332a780eb6d4-var-log-ovn\") pod \"ovn-controller-tb2c4-config-gb9jj\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " pod="openstack/ovn-controller-tb2c4-config-gb9jj" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.681016 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/49f6d2fc-6fd2-4823-befc-332a780eb6d4-additional-scripts\") pod \"ovn-controller-tb2c4-config-gb9jj\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " pod="openstack/ovn-controller-tb2c4-config-gb9jj" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.782362 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49f6d2fc-6fd2-4823-befc-332a780eb6d4-scripts\") pod \"ovn-controller-tb2c4-config-gb9jj\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " pod="openstack/ovn-controller-tb2c4-config-gb9jj" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.782401 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/49f6d2fc-6fd2-4823-befc-332a780eb6d4-var-run\") pod \"ovn-controller-tb2c4-config-gb9jj\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " pod="openstack/ovn-controller-tb2c4-config-gb9jj" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.782427 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7v9zf\" (UniqueName: \"kubernetes.io/projected/49f6d2fc-6fd2-4823-befc-332a780eb6d4-kube-api-access-7v9zf\") pod \"ovn-controller-tb2c4-config-gb9jj\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " pod="openstack/ovn-controller-tb2c4-config-gb9jj" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.782453 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/49f6d2fc-6fd2-4823-befc-332a780eb6d4-var-run-ovn\") pod \"ovn-controller-tb2c4-config-gb9jj\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " pod="openstack/ovn-controller-tb2c4-config-gb9jj" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.782477 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/49f6d2fc-6fd2-4823-befc-332a780eb6d4-var-log-ovn\") pod \"ovn-controller-tb2c4-config-gb9jj\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " pod="openstack/ovn-controller-tb2c4-config-gb9jj" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.782514 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/49f6d2fc-6fd2-4823-befc-332a780eb6d4-additional-scripts\") pod \"ovn-controller-tb2c4-config-gb9jj\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " pod="openstack/ovn-controller-tb2c4-config-gb9jj" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.782804 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/49f6d2fc-6fd2-4823-befc-332a780eb6d4-var-run\") pod \"ovn-controller-tb2c4-config-gb9jj\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " pod="openstack/ovn-controller-tb2c4-config-gb9jj" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.782809 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/49f6d2fc-6fd2-4823-befc-332a780eb6d4-var-run-ovn\") pod \"ovn-controller-tb2c4-config-gb9jj\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " pod="openstack/ovn-controller-tb2c4-config-gb9jj" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.782905 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/49f6d2fc-6fd2-4823-befc-332a780eb6d4-var-log-ovn\") pod \"ovn-controller-tb2c4-config-gb9jj\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " pod="openstack/ovn-controller-tb2c4-config-gb9jj" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.783363 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/49f6d2fc-6fd2-4823-befc-332a780eb6d4-additional-scripts\") pod \"ovn-controller-tb2c4-config-gb9jj\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " pod="openstack/ovn-controller-tb2c4-config-gb9jj" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.784825 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49f6d2fc-6fd2-4823-befc-332a780eb6d4-scripts\") pod \"ovn-controller-tb2c4-config-gb9jj\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " pod="openstack/ovn-controller-tb2c4-config-gb9jj" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.804830 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7v9zf\" (UniqueName: \"kubernetes.io/projected/49f6d2fc-6fd2-4823-befc-332a780eb6d4-kube-api-access-7v9zf\") pod \"ovn-controller-tb2c4-config-gb9jj\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " pod="openstack/ovn-controller-tb2c4-config-gb9jj" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.895842 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tb2c4-config-gb9jj" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.923390 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-tb2c4" Oct 13 09:01:04 crc kubenswrapper[4685]: I1013 09:01:04.955152 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.006101 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.459958 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-f2mfh"] Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.461246 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-f2mfh" Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.470204 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-f2mfh"] Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.528633 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd819379-6804-4352-8ad0-416859783cd4" path="/var/lib/kubelet/pods/cd819379-6804-4352-8ad0-416859783cd4/volumes" Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.545254 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-dp86g"] Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.546806 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dp86g" Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.597898 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4k4x\" (UniqueName: \"kubernetes.io/projected/6870aa8c-c1e3-4c3c-b1dd-c737e0a1d382-kube-api-access-m4k4x\") pod \"cinder-db-create-f2mfh\" (UID: \"6870aa8c-c1e3-4c3c-b1dd-c737e0a1d382\") " pod="openstack/cinder-db-create-f2mfh" Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.627977 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-dp86g"] Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.672851 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tb2c4-config-gb9jj"] Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.700744 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5fkc\" (UniqueName: \"kubernetes.io/projected/aab3ac63-a140-47f9-9fa3-113d412dd2d5-kube-api-access-g5fkc\") pod \"barbican-db-create-dp86g\" (UID: \"aab3ac63-a140-47f9-9fa3-113d412dd2d5\") " pod="openstack/barbican-db-create-dp86g" Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.700833 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4k4x\" (UniqueName: \"kubernetes.io/projected/6870aa8c-c1e3-4c3c-b1dd-c737e0a1d382-kube-api-access-m4k4x\") pod \"cinder-db-create-f2mfh\" (UID: \"6870aa8c-c1e3-4c3c-b1dd-c737e0a1d382\") " pod="openstack/cinder-db-create-f2mfh" Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.768530 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4k4x\" (UniqueName: \"kubernetes.io/projected/6870aa8c-c1e3-4c3c-b1dd-c737e0a1d382-kube-api-access-m4k4x\") pod \"cinder-db-create-f2mfh\" (UID: \"6870aa8c-c1e3-4c3c-b1dd-c737e0a1d382\") " pod="openstack/cinder-db-create-f2mfh" Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.781579 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-l25fj"] Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.783367 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-l25fj" Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.787814 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.788056 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.788115 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5xp9z" Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.795428 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.805305 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5fkc\" (UniqueName: \"kubernetes.io/projected/aab3ac63-a140-47f9-9fa3-113d412dd2d5-kube-api-access-g5fkc\") pod \"barbican-db-create-dp86g\" (UID: \"aab3ac63-a140-47f9-9fa3-113d412dd2d5\") " pod="openstack/barbican-db-create-dp86g" Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.827090 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-l25fj"] Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.827582 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-f2mfh" Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.843668 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5fkc\" (UniqueName: \"kubernetes.io/projected/aab3ac63-a140-47f9-9fa3-113d412dd2d5-kube-api-access-g5fkc\") pod \"barbican-db-create-dp86g\" (UID: \"aab3ac63-a140-47f9-9fa3-113d412dd2d5\") " pod="openstack/barbican-db-create-dp86g" Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.855030 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-87wjr"] Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.858004 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-87wjr" Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.877962 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dp86g" Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.893337 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-87wjr"] Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.916478 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77gkz\" (UniqueName: \"kubernetes.io/projected/566d7edf-ec33-43e8-abd4-7e180a2fda57-kube-api-access-77gkz\") pod \"keystone-db-sync-l25fj\" (UID: \"566d7edf-ec33-43e8-abd4-7e180a2fda57\") " pod="openstack/keystone-db-sync-l25fj" Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.917018 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/566d7edf-ec33-43e8-abd4-7e180a2fda57-combined-ca-bundle\") pod \"keystone-db-sync-l25fj\" (UID: \"566d7edf-ec33-43e8-abd4-7e180a2fda57\") " pod="openstack/keystone-db-sync-l25fj" Oct 13 09:01:05 crc kubenswrapper[4685]: I1013 09:01:05.917097 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/566d7edf-ec33-43e8-abd4-7e180a2fda57-config-data\") pod \"keystone-db-sync-l25fj\" (UID: \"566d7edf-ec33-43e8-abd4-7e180a2fda57\") " pod="openstack/keystone-db-sync-l25fj" Oct 13 09:01:06 crc kubenswrapper[4685]: I1013 09:01:06.019426 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mz2c\" (UniqueName: \"kubernetes.io/projected/c201a91a-9c59-46e8-bd69-cde6bd87d14a-kube-api-access-8mz2c\") pod \"neutron-db-create-87wjr\" (UID: \"c201a91a-9c59-46e8-bd69-cde6bd87d14a\") " pod="openstack/neutron-db-create-87wjr" Oct 13 09:01:06 crc kubenswrapper[4685]: I1013 09:01:06.019524 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77gkz\" (UniqueName: \"kubernetes.io/projected/566d7edf-ec33-43e8-abd4-7e180a2fda57-kube-api-access-77gkz\") pod \"keystone-db-sync-l25fj\" (UID: \"566d7edf-ec33-43e8-abd4-7e180a2fda57\") " pod="openstack/keystone-db-sync-l25fj" Oct 13 09:01:06 crc kubenswrapper[4685]: I1013 09:01:06.019573 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/566d7edf-ec33-43e8-abd4-7e180a2fda57-combined-ca-bundle\") pod \"keystone-db-sync-l25fj\" (UID: \"566d7edf-ec33-43e8-abd4-7e180a2fda57\") " pod="openstack/keystone-db-sync-l25fj" Oct 13 09:01:06 crc kubenswrapper[4685]: I1013 09:01:06.020456 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/566d7edf-ec33-43e8-abd4-7e180a2fda57-config-data\") pod \"keystone-db-sync-l25fj\" (UID: \"566d7edf-ec33-43e8-abd4-7e180a2fda57\") " pod="openstack/keystone-db-sync-l25fj" Oct 13 09:01:06 crc kubenswrapper[4685]: I1013 09:01:06.028856 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/566d7edf-ec33-43e8-abd4-7e180a2fda57-combined-ca-bundle\") pod \"keystone-db-sync-l25fj\" (UID: \"566d7edf-ec33-43e8-abd4-7e180a2fda57\") " pod="openstack/keystone-db-sync-l25fj" Oct 13 09:01:06 crc kubenswrapper[4685]: I1013 09:01:06.032871 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/566d7edf-ec33-43e8-abd4-7e180a2fda57-config-data\") pod \"keystone-db-sync-l25fj\" (UID: \"566d7edf-ec33-43e8-abd4-7e180a2fda57\") " pod="openstack/keystone-db-sync-l25fj" Oct 13 09:01:06 crc kubenswrapper[4685]: I1013 09:01:06.043300 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77gkz\" (UniqueName: \"kubernetes.io/projected/566d7edf-ec33-43e8-abd4-7e180a2fda57-kube-api-access-77gkz\") pod \"keystone-db-sync-l25fj\" (UID: \"566d7edf-ec33-43e8-abd4-7e180a2fda57\") " pod="openstack/keystone-db-sync-l25fj" Oct 13 09:01:06 crc kubenswrapper[4685]: I1013 09:01:06.126557 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mz2c\" (UniqueName: \"kubernetes.io/projected/c201a91a-9c59-46e8-bd69-cde6bd87d14a-kube-api-access-8mz2c\") pod \"neutron-db-create-87wjr\" (UID: \"c201a91a-9c59-46e8-bd69-cde6bd87d14a\") " pod="openstack/neutron-db-create-87wjr" Oct 13 09:01:06 crc kubenswrapper[4685]: I1013 09:01:06.127147 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-l25fj" Oct 13 09:01:06 crc kubenswrapper[4685]: I1013 09:01:06.143949 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mz2c\" (UniqueName: \"kubernetes.io/projected/c201a91a-9c59-46e8-bd69-cde6bd87d14a-kube-api-access-8mz2c\") pod \"neutron-db-create-87wjr\" (UID: \"c201a91a-9c59-46e8-bd69-cde6bd87d14a\") " pod="openstack/neutron-db-create-87wjr" Oct 13 09:01:06 crc kubenswrapper[4685]: I1013 09:01:06.154874 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tb2c4-config-gb9jj" event={"ID":"49f6d2fc-6fd2-4823-befc-332a780eb6d4","Type":"ContainerStarted","Data":"14ea6266ec1b5d194aa959ff938ecccfd535274b2b325f72f422f5261545c958"} Oct 13 09:01:06 crc kubenswrapper[4685]: I1013 09:01:06.230986 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-87wjr" Oct 13 09:01:06 crc kubenswrapper[4685]: I1013 09:01:06.503905 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-dp86g"] Oct 13 09:01:06 crc kubenswrapper[4685]: I1013 09:01:06.656431 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-f2mfh"] Oct 13 09:01:06 crc kubenswrapper[4685]: I1013 09:01:06.759415 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-l25fj"] Oct 13 09:01:06 crc kubenswrapper[4685]: W1013 09:01:06.779083 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod566d7edf_ec33_43e8_abd4_7e180a2fda57.slice/crio-d098c27e5fe3d9863e43ac51e01ad1e6ead01110b163c07807101bc762fa1671 WatchSource:0}: Error finding container d098c27e5fe3d9863e43ac51e01ad1e6ead01110b163c07807101bc762fa1671: Status 404 returned error can't find the container with id d098c27e5fe3d9863e43ac51e01ad1e6ead01110b163c07807101bc762fa1671 Oct 13 09:01:06 crc kubenswrapper[4685]: I1013 09:01:06.833105 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-87wjr"] Oct 13 09:01:07 crc kubenswrapper[4685]: I1013 09:01:07.164089 4685 generic.go:334] "Generic (PLEG): container finished" podID="6870aa8c-c1e3-4c3c-b1dd-c737e0a1d382" containerID="933c94f4464c94619faea471b44761d3cffdd20a36877e83f316467606fbf645" exitCode=0 Oct 13 09:01:07 crc kubenswrapper[4685]: I1013 09:01:07.164235 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-f2mfh" event={"ID":"6870aa8c-c1e3-4c3c-b1dd-c737e0a1d382","Type":"ContainerDied","Data":"933c94f4464c94619faea471b44761d3cffdd20a36877e83f316467606fbf645"} Oct 13 09:01:07 crc kubenswrapper[4685]: I1013 09:01:07.164359 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-f2mfh" event={"ID":"6870aa8c-c1e3-4c3c-b1dd-c737e0a1d382","Type":"ContainerStarted","Data":"5cd84d8833e22a936f1100ce3f0d6370d25843db5b8b0fa62dcd117487f17bba"} Oct 13 09:01:07 crc kubenswrapper[4685]: I1013 09:01:07.168623 4685 generic.go:334] "Generic (PLEG): container finished" podID="49f6d2fc-6fd2-4823-befc-332a780eb6d4" containerID="0df2a8d7eb43f1a38cf4e1b537678a2c97537c359396ac708b4e4ca59a9b58cf" exitCode=0 Oct 13 09:01:07 crc kubenswrapper[4685]: I1013 09:01:07.168680 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tb2c4-config-gb9jj" event={"ID":"49f6d2fc-6fd2-4823-befc-332a780eb6d4","Type":"ContainerDied","Data":"0df2a8d7eb43f1a38cf4e1b537678a2c97537c359396ac708b4e4ca59a9b58cf"} Oct 13 09:01:07 crc kubenswrapper[4685]: I1013 09:01:07.172492 4685 generic.go:334] "Generic (PLEG): container finished" podID="aab3ac63-a140-47f9-9fa3-113d412dd2d5" containerID="6a37957315414d8b9f2fefba688c4bcab605c61cef7b622d0f1955d63cc3d202" exitCode=0 Oct 13 09:01:07 crc kubenswrapper[4685]: I1013 09:01:07.172561 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dp86g" event={"ID":"aab3ac63-a140-47f9-9fa3-113d412dd2d5","Type":"ContainerDied","Data":"6a37957315414d8b9f2fefba688c4bcab605c61cef7b622d0f1955d63cc3d202"} Oct 13 09:01:07 crc kubenswrapper[4685]: I1013 09:01:07.172589 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dp86g" event={"ID":"aab3ac63-a140-47f9-9fa3-113d412dd2d5","Type":"ContainerStarted","Data":"b87adaba17916577e42d54ae600bdb287019573a6d60ef0949dd30b8132d345e"} Oct 13 09:01:07 crc kubenswrapper[4685]: I1013 09:01:07.174729 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-87wjr" event={"ID":"c201a91a-9c59-46e8-bd69-cde6bd87d14a","Type":"ContainerStarted","Data":"f19f44d57a3ffc94c5ddeee385aa1465ae5df8bc8fddcd326065453d05a59db4"} Oct 13 09:01:07 crc kubenswrapper[4685]: I1013 09:01:07.176276 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-l25fj" event={"ID":"566d7edf-ec33-43e8-abd4-7e180a2fda57","Type":"ContainerStarted","Data":"d098c27e5fe3d9863e43ac51e01ad1e6ead01110b163c07807101bc762fa1671"} Oct 13 09:01:12 crc kubenswrapper[4685]: I1013 09:01:12.949749 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-etc-swift\") pod \"swift-storage-0\" (UID: \"0791156b-11f2-43ad-b910-3e42b4d6670f\") " pod="openstack/swift-storage-0" Oct 13 09:01:12 crc kubenswrapper[4685]: I1013 09:01:12.972210 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0791156b-11f2-43ad-b910-3e42b4d6670f-etc-swift\") pod \"swift-storage-0\" (UID: \"0791156b-11f2-43ad-b910-3e42b4d6670f\") " pod="openstack/swift-storage-0" Oct 13 09:01:13 crc kubenswrapper[4685]: I1013 09:01:13.227555 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 13 09:01:14 crc kubenswrapper[4685]: I1013 09:01:14.916401 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tb2c4-config-gb9jj" Oct 13 09:01:14 crc kubenswrapper[4685]: I1013 09:01:14.923644 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dp86g" Oct 13 09:01:14 crc kubenswrapper[4685]: I1013 09:01:14.924088 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-f2mfh" Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.084819 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/49f6d2fc-6fd2-4823-befc-332a780eb6d4-var-log-ovn\") pod \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.085170 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/49f6d2fc-6fd2-4823-befc-332a780eb6d4-var-run-ovn\") pod \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.085232 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/49f6d2fc-6fd2-4823-befc-332a780eb6d4-var-run\") pod \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.085279 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7v9zf\" (UniqueName: \"kubernetes.io/projected/49f6d2fc-6fd2-4823-befc-332a780eb6d4-kube-api-access-7v9zf\") pod \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.085331 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4k4x\" (UniqueName: \"kubernetes.io/projected/6870aa8c-c1e3-4c3c-b1dd-c737e0a1d382-kube-api-access-m4k4x\") pod \"6870aa8c-c1e3-4c3c-b1dd-c737e0a1d382\" (UID: \"6870aa8c-c1e3-4c3c-b1dd-c737e0a1d382\") " Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.085428 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/49f6d2fc-6fd2-4823-befc-332a780eb6d4-additional-scripts\") pod \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.085513 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49f6d2fc-6fd2-4823-befc-332a780eb6d4-scripts\") pod \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\" (UID: \"49f6d2fc-6fd2-4823-befc-332a780eb6d4\") " Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.085561 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5fkc\" (UniqueName: \"kubernetes.io/projected/aab3ac63-a140-47f9-9fa3-113d412dd2d5-kube-api-access-g5fkc\") pod \"aab3ac63-a140-47f9-9fa3-113d412dd2d5\" (UID: \"aab3ac63-a140-47f9-9fa3-113d412dd2d5\") " Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.084950 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/49f6d2fc-6fd2-4823-befc-332a780eb6d4-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "49f6d2fc-6fd2-4823-befc-332a780eb6d4" (UID: "49f6d2fc-6fd2-4823-befc-332a780eb6d4"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.086366 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/49f6d2fc-6fd2-4823-befc-332a780eb6d4-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "49f6d2fc-6fd2-4823-befc-332a780eb6d4" (UID: "49f6d2fc-6fd2-4823-befc-332a780eb6d4"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.086388 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/49f6d2fc-6fd2-4823-befc-332a780eb6d4-var-run" (OuterVolumeSpecName: "var-run") pod "49f6d2fc-6fd2-4823-befc-332a780eb6d4" (UID: "49f6d2fc-6fd2-4823-befc-332a780eb6d4"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.086580 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49f6d2fc-6fd2-4823-befc-332a780eb6d4-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "49f6d2fc-6fd2-4823-befc-332a780eb6d4" (UID: "49f6d2fc-6fd2-4823-befc-332a780eb6d4"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.086812 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49f6d2fc-6fd2-4823-befc-332a780eb6d4-scripts" (OuterVolumeSpecName: "scripts") pod "49f6d2fc-6fd2-4823-befc-332a780eb6d4" (UID: "49f6d2fc-6fd2-4823-befc-332a780eb6d4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.089896 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aab3ac63-a140-47f9-9fa3-113d412dd2d5-kube-api-access-g5fkc" (OuterVolumeSpecName: "kube-api-access-g5fkc") pod "aab3ac63-a140-47f9-9fa3-113d412dd2d5" (UID: "aab3ac63-a140-47f9-9fa3-113d412dd2d5"). InnerVolumeSpecName "kube-api-access-g5fkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.090177 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6870aa8c-c1e3-4c3c-b1dd-c737e0a1d382-kube-api-access-m4k4x" (OuterVolumeSpecName: "kube-api-access-m4k4x") pod "6870aa8c-c1e3-4c3c-b1dd-c737e0a1d382" (UID: "6870aa8c-c1e3-4c3c-b1dd-c737e0a1d382"). InnerVolumeSpecName "kube-api-access-m4k4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.090411 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49f6d2fc-6fd2-4823-befc-332a780eb6d4-kube-api-access-7v9zf" (OuterVolumeSpecName: "kube-api-access-7v9zf") pod "49f6d2fc-6fd2-4823-befc-332a780eb6d4" (UID: "49f6d2fc-6fd2-4823-befc-332a780eb6d4"). InnerVolumeSpecName "kube-api-access-7v9zf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.189403 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7v9zf\" (UniqueName: \"kubernetes.io/projected/49f6d2fc-6fd2-4823-befc-332a780eb6d4-kube-api-access-7v9zf\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.190049 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4k4x\" (UniqueName: \"kubernetes.io/projected/6870aa8c-c1e3-4c3c-b1dd-c737e0a1d382-kube-api-access-m4k4x\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.190078 4685 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/49f6d2fc-6fd2-4823-befc-332a780eb6d4-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.190092 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49f6d2fc-6fd2-4823-befc-332a780eb6d4-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.190102 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5fkc\" (UniqueName: \"kubernetes.io/projected/aab3ac63-a140-47f9-9fa3-113d412dd2d5-kube-api-access-g5fkc\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.190112 4685 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/49f6d2fc-6fd2-4823-befc-332a780eb6d4-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.190121 4685 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/49f6d2fc-6fd2-4823-befc-332a780eb6d4-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.190130 4685 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/49f6d2fc-6fd2-4823-befc-332a780eb6d4-var-run\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.249786 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 13 09:01:15 crc kubenswrapper[4685]: W1013 09:01:15.253749 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0791156b_11f2_43ad_b910_3e42b4d6670f.slice/crio-c8df2368d7cd1bb34f45f4658956fb5c959603231e324400c458cc119cb2f6c6 WatchSource:0}: Error finding container c8df2368d7cd1bb34f45f4658956fb5c959603231e324400c458cc119cb2f6c6: Status 404 returned error can't find the container with id c8df2368d7cd1bb34f45f4658956fb5c959603231e324400c458cc119cb2f6c6 Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.257315 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tb2c4-config-gb9jj" event={"ID":"49f6d2fc-6fd2-4823-befc-332a780eb6d4","Type":"ContainerDied","Data":"14ea6266ec1b5d194aa959ff938ecccfd535274b2b325f72f422f5261545c958"} Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.257397 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14ea6266ec1b5d194aa959ff938ecccfd535274b2b325f72f422f5261545c958" Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.257330 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tb2c4-config-gb9jj" Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.263271 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dp86g" event={"ID":"aab3ac63-a140-47f9-9fa3-113d412dd2d5","Type":"ContainerDied","Data":"b87adaba17916577e42d54ae600bdb287019573a6d60ef0949dd30b8132d345e"} Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.263301 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dp86g" Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.263314 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b87adaba17916577e42d54ae600bdb287019573a6d60ef0949dd30b8132d345e" Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.266497 4685 generic.go:334] "Generic (PLEG): container finished" podID="c201a91a-9c59-46e8-bd69-cde6bd87d14a" containerID="a7425caa7079a13a6b62cc8bbdeb7f3b30af1b8418964a6473386ca4c5d7d855" exitCode=0 Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.266575 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-87wjr" event={"ID":"c201a91a-9c59-46e8-bd69-cde6bd87d14a","Type":"ContainerDied","Data":"a7425caa7079a13a6b62cc8bbdeb7f3b30af1b8418964a6473386ca4c5d7d855"} Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.269541 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-f2mfh" event={"ID":"6870aa8c-c1e3-4c3c-b1dd-c737e0a1d382","Type":"ContainerDied","Data":"5cd84d8833e22a936f1100ce3f0d6370d25843db5b8b0fa62dcd117487f17bba"} Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.269574 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5cd84d8833e22a936f1100ce3f0d6370d25843db5b8b0fa62dcd117487f17bba" Oct 13 09:01:15 crc kubenswrapper[4685]: I1013 09:01:15.269593 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-f2mfh" Oct 13 09:01:16 crc kubenswrapper[4685]: I1013 09:01:16.009325 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-tb2c4-config-gb9jj"] Oct 13 09:01:16 crc kubenswrapper[4685]: I1013 09:01:16.019516 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-tb2c4-config-gb9jj"] Oct 13 09:01:16 crc kubenswrapper[4685]: I1013 09:01:16.287183 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0791156b-11f2-43ad-b910-3e42b4d6670f","Type":"ContainerStarted","Data":"c8df2368d7cd1bb34f45f4658956fb5c959603231e324400c458cc119cb2f6c6"} Oct 13 09:01:16 crc kubenswrapper[4685]: I1013 09:01:16.290104 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jvxxs" event={"ID":"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251","Type":"ContainerStarted","Data":"3f7fb51165e65e09f0de8b115d50f44b4dca84584b4d149722bbd704aa2a4718"} Oct 13 09:01:16 crc kubenswrapper[4685]: I1013 09:01:16.315969 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-jvxxs" podStartSLOduration=2.88153575 podStartE2EDuration="18.315950307s" podCreationTimestamp="2025-10-13 09:00:58 +0000 UTC" firstStartedPulling="2025-10-13 09:00:59.349542944 +0000 UTC m=+984.497418705" lastFinishedPulling="2025-10-13 09:01:14.783957501 +0000 UTC m=+999.931833262" observedRunningTime="2025-10-13 09:01:16.311034574 +0000 UTC m=+1001.458910335" watchObservedRunningTime="2025-10-13 09:01:16.315950307 +0000 UTC m=+1001.463826308" Oct 13 09:01:17 crc kubenswrapper[4685]: I1013 09:01:17.514200 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49f6d2fc-6fd2-4823-befc-332a780eb6d4" path="/var/lib/kubelet/pods/49f6d2fc-6fd2-4823-befc-332a780eb6d4/volumes" Oct 13 09:01:19 crc kubenswrapper[4685]: I1013 09:01:19.193022 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-87wjr" Oct 13 09:01:19 crc kubenswrapper[4685]: I1013 09:01:19.331632 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-87wjr" event={"ID":"c201a91a-9c59-46e8-bd69-cde6bd87d14a","Type":"ContainerDied","Data":"f19f44d57a3ffc94c5ddeee385aa1465ae5df8bc8fddcd326065453d05a59db4"} Oct 13 09:01:19 crc kubenswrapper[4685]: I1013 09:01:19.331673 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f19f44d57a3ffc94c5ddeee385aa1465ae5df8bc8fddcd326065453d05a59db4" Oct 13 09:01:19 crc kubenswrapper[4685]: I1013 09:01:19.331706 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-87wjr" Oct 13 09:01:19 crc kubenswrapper[4685]: I1013 09:01:19.352583 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mz2c\" (UniqueName: \"kubernetes.io/projected/c201a91a-9c59-46e8-bd69-cde6bd87d14a-kube-api-access-8mz2c\") pod \"c201a91a-9c59-46e8-bd69-cde6bd87d14a\" (UID: \"c201a91a-9c59-46e8-bd69-cde6bd87d14a\") " Oct 13 09:01:19 crc kubenswrapper[4685]: I1013 09:01:19.356617 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c201a91a-9c59-46e8-bd69-cde6bd87d14a-kube-api-access-8mz2c" (OuterVolumeSpecName: "kube-api-access-8mz2c") pod "c201a91a-9c59-46e8-bd69-cde6bd87d14a" (UID: "c201a91a-9c59-46e8-bd69-cde6bd87d14a"). InnerVolumeSpecName "kube-api-access-8mz2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:01:19 crc kubenswrapper[4685]: I1013 09:01:19.454798 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mz2c\" (UniqueName: \"kubernetes.io/projected/c201a91a-9c59-46e8-bd69-cde6bd87d14a-kube-api-access-8mz2c\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:20 crc kubenswrapper[4685]: I1013 09:01:20.344398 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-l25fj" event={"ID":"566d7edf-ec33-43e8-abd4-7e180a2fda57","Type":"ContainerStarted","Data":"2620c9584b85313a1c833d0dce417aec96153ad94723a334d24b54af9ba9ce20"} Oct 13 09:01:20 crc kubenswrapper[4685]: I1013 09:01:20.348112 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0791156b-11f2-43ad-b910-3e42b4d6670f","Type":"ContainerStarted","Data":"b7b8accebc7124930cf4b55f7a6547f7dc240bcda41846b68140c6cbc504f01f"} Oct 13 09:01:20 crc kubenswrapper[4685]: I1013 09:01:20.348167 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0791156b-11f2-43ad-b910-3e42b4d6670f","Type":"ContainerStarted","Data":"c7d6a99f4592f6f87d6e4d8080ec76f209fdeb03d534972362aea738bb59bf00"} Oct 13 09:01:20 crc kubenswrapper[4685]: I1013 09:01:20.390794 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-l25fj" podStartSLOduration=2.961521606 podStartE2EDuration="15.390775734s" podCreationTimestamp="2025-10-13 09:01:05 +0000 UTC" firstStartedPulling="2025-10-13 09:01:06.781304031 +0000 UTC m=+991.929179792" lastFinishedPulling="2025-10-13 09:01:19.210558159 +0000 UTC m=+1004.358433920" observedRunningTime="2025-10-13 09:01:20.382053136 +0000 UTC m=+1005.529928897" watchObservedRunningTime="2025-10-13 09:01:20.390775734 +0000 UTC m=+1005.538651495" Oct 13 09:01:21 crc kubenswrapper[4685]: I1013 09:01:21.359992 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0791156b-11f2-43ad-b910-3e42b4d6670f","Type":"ContainerStarted","Data":"d031e458089fa63229ac9061c4b422e2dbf463f8e190debe596bb56913029f70"} Oct 13 09:01:21 crc kubenswrapper[4685]: I1013 09:01:21.361011 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0791156b-11f2-43ad-b910-3e42b4d6670f","Type":"ContainerStarted","Data":"064392e8d781e766b650dff2cef8f7b9279f1ed460d75ff6585664d4cfa1a9d0"} Oct 13 09:01:22 crc kubenswrapper[4685]: I1013 09:01:22.368824 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0791156b-11f2-43ad-b910-3e42b4d6670f","Type":"ContainerStarted","Data":"6664377112c5c242bb7e74223c3aa13b399c1e4c335527d83f722fe3ecf9cb1e"} Oct 13 09:01:22 crc kubenswrapper[4685]: I1013 09:01:22.369102 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0791156b-11f2-43ad-b910-3e42b4d6670f","Type":"ContainerStarted","Data":"c60a2b6258d0f9e484871155835de203a3c7f76e1747473c87dfbc882c38a47d"} Oct 13 09:01:22 crc kubenswrapper[4685]: I1013 09:01:22.369113 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0791156b-11f2-43ad-b910-3e42b4d6670f","Type":"ContainerStarted","Data":"71729abcc3623f0d2578789b661f236f145b8009f10cfda342ca42c40c077475"} Oct 13 09:01:22 crc kubenswrapper[4685]: I1013 09:01:22.980102 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:01:22 crc kubenswrapper[4685]: I1013 09:01:22.980448 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:01:23 crc kubenswrapper[4685]: I1013 09:01:23.379116 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0791156b-11f2-43ad-b910-3e42b4d6670f","Type":"ContainerStarted","Data":"5e29735dd0170703cc54416c40b1b9129da68bc004270a0fc085001301ead64e"} Oct 13 09:01:23 crc kubenswrapper[4685]: I1013 09:01:23.380666 4685 generic.go:334] "Generic (PLEG): container finished" podID="566d7edf-ec33-43e8-abd4-7e180a2fda57" containerID="2620c9584b85313a1c833d0dce417aec96153ad94723a334d24b54af9ba9ce20" exitCode=0 Oct 13 09:01:23 crc kubenswrapper[4685]: I1013 09:01:23.380703 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-l25fj" event={"ID":"566d7edf-ec33-43e8-abd4-7e180a2fda57","Type":"ContainerDied","Data":"2620c9584b85313a1c833d0dce417aec96153ad94723a334d24b54af9ba9ce20"} Oct 13 09:01:24 crc kubenswrapper[4685]: I1013 09:01:24.394972 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0791156b-11f2-43ad-b910-3e42b4d6670f","Type":"ContainerStarted","Data":"b7b139c5c404e0539289357e8e597ddd43db3b4a6a4dd6b2f3cf6fc6fc202456"} Oct 13 09:01:24 crc kubenswrapper[4685]: I1013 09:01:24.395329 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0791156b-11f2-43ad-b910-3e42b4d6670f","Type":"ContainerStarted","Data":"c5cbfe3dccc5c8a60eb76f8d6bf2348cf190cf7960e69f48cfd34cc458bd74b9"} Oct 13 09:01:24 crc kubenswrapper[4685]: I1013 09:01:24.395351 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0791156b-11f2-43ad-b910-3e42b4d6670f","Type":"ContainerStarted","Data":"552e6739bc6e1403e0630452bdc50ef9dbac6323ec8e8fe9349d6e3eb63b8257"} Oct 13 09:01:24 crc kubenswrapper[4685]: I1013 09:01:24.395371 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0791156b-11f2-43ad-b910-3e42b4d6670f","Type":"ContainerStarted","Data":"ee517d30e18c34897ff2c7376b002b5eaba935e3b1dead9d6dbe80fa597a298b"} Oct 13 09:01:24 crc kubenswrapper[4685]: I1013 09:01:24.715448 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-l25fj" Oct 13 09:01:24 crc kubenswrapper[4685]: I1013 09:01:24.861722 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/566d7edf-ec33-43e8-abd4-7e180a2fda57-config-data\") pod \"566d7edf-ec33-43e8-abd4-7e180a2fda57\" (UID: \"566d7edf-ec33-43e8-abd4-7e180a2fda57\") " Oct 13 09:01:24 crc kubenswrapper[4685]: I1013 09:01:24.861802 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/566d7edf-ec33-43e8-abd4-7e180a2fda57-combined-ca-bundle\") pod \"566d7edf-ec33-43e8-abd4-7e180a2fda57\" (UID: \"566d7edf-ec33-43e8-abd4-7e180a2fda57\") " Oct 13 09:01:24 crc kubenswrapper[4685]: I1013 09:01:24.861963 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77gkz\" (UniqueName: \"kubernetes.io/projected/566d7edf-ec33-43e8-abd4-7e180a2fda57-kube-api-access-77gkz\") pod \"566d7edf-ec33-43e8-abd4-7e180a2fda57\" (UID: \"566d7edf-ec33-43e8-abd4-7e180a2fda57\") " Oct 13 09:01:24 crc kubenswrapper[4685]: I1013 09:01:24.867733 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/566d7edf-ec33-43e8-abd4-7e180a2fda57-kube-api-access-77gkz" (OuterVolumeSpecName: "kube-api-access-77gkz") pod "566d7edf-ec33-43e8-abd4-7e180a2fda57" (UID: "566d7edf-ec33-43e8-abd4-7e180a2fda57"). InnerVolumeSpecName "kube-api-access-77gkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:01:24 crc kubenswrapper[4685]: I1013 09:01:24.965495 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77gkz\" (UniqueName: \"kubernetes.io/projected/566d7edf-ec33-43e8-abd4-7e180a2fda57-kube-api-access-77gkz\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:24 crc kubenswrapper[4685]: I1013 09:01:24.965621 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/566d7edf-ec33-43e8-abd4-7e180a2fda57-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "566d7edf-ec33-43e8-abd4-7e180a2fda57" (UID: "566d7edf-ec33-43e8-abd4-7e180a2fda57"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.002785 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/566d7edf-ec33-43e8-abd4-7e180a2fda57-config-data" (OuterVolumeSpecName: "config-data") pod "566d7edf-ec33-43e8-abd4-7e180a2fda57" (UID: "566d7edf-ec33-43e8-abd4-7e180a2fda57"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.067092 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/566d7edf-ec33-43e8-abd4-7e180a2fda57-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.067127 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/566d7edf-ec33-43e8-abd4-7e180a2fda57-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.406690 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-l25fj" event={"ID":"566d7edf-ec33-43e8-abd4-7e180a2fda57","Type":"ContainerDied","Data":"d098c27e5fe3d9863e43ac51e01ad1e6ead01110b163c07807101bc762fa1671"} Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.406756 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d098c27e5fe3d9863e43ac51e01ad1e6ead01110b163c07807101bc762fa1671" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.406843 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-l25fj" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.418871 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0791156b-11f2-43ad-b910-3e42b4d6670f","Type":"ContainerStarted","Data":"7d96d7d63c7426198d7b276bc38c52a812a5863495ffe6e75a6f02aaf2708715"} Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.418963 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0791156b-11f2-43ad-b910-3e42b4d6670f","Type":"ContainerStarted","Data":"7036cd8669d27d81ca3b67637be840b3583d30e5bc51595bc4d1be49635b8dc1"} Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.418974 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0791156b-11f2-43ad-b910-3e42b4d6670f","Type":"ContainerStarted","Data":"8942e823e7ddc25ec6bf98adf19d5eef599dcc4cb5a4a9fd6d88ad2f647ae5f3"} Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.458407 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=38.252954184000004 podStartE2EDuration="46.458392449s" podCreationTimestamp="2025-10-13 09:00:39 +0000 UTC" firstStartedPulling="2025-10-13 09:01:15.256871209 +0000 UTC m=+1000.404746970" lastFinishedPulling="2025-10-13 09:01:23.462309464 +0000 UTC m=+1008.610185235" observedRunningTime="2025-10-13 09:01:25.455550892 +0000 UTC m=+1010.603426663" watchObservedRunningTime="2025-10-13 09:01:25.458392449 +0000 UTC m=+1010.606268210" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.516904 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-2261-account-create-ndcwb"] Oct 13 09:01:25 crc kubenswrapper[4685]: E1013 09:01:25.517184 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6870aa8c-c1e3-4c3c-b1dd-c737e0a1d382" containerName="mariadb-database-create" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.517199 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6870aa8c-c1e3-4c3c-b1dd-c737e0a1d382" containerName="mariadb-database-create" Oct 13 09:01:25 crc kubenswrapper[4685]: E1013 09:01:25.517212 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49f6d2fc-6fd2-4823-befc-332a780eb6d4" containerName="ovn-config" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.517220 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="49f6d2fc-6fd2-4823-befc-332a780eb6d4" containerName="ovn-config" Oct 13 09:01:25 crc kubenswrapper[4685]: E1013 09:01:25.517237 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c201a91a-9c59-46e8-bd69-cde6bd87d14a" containerName="mariadb-database-create" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.517243 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c201a91a-9c59-46e8-bd69-cde6bd87d14a" containerName="mariadb-database-create" Oct 13 09:01:25 crc kubenswrapper[4685]: E1013 09:01:25.517260 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aab3ac63-a140-47f9-9fa3-113d412dd2d5" containerName="mariadb-database-create" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.517265 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="aab3ac63-a140-47f9-9fa3-113d412dd2d5" containerName="mariadb-database-create" Oct 13 09:01:25 crc kubenswrapper[4685]: E1013 09:01:25.517275 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="566d7edf-ec33-43e8-abd4-7e180a2fda57" containerName="keystone-db-sync" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.517280 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="566d7edf-ec33-43e8-abd4-7e180a2fda57" containerName="keystone-db-sync" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.517451 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="c201a91a-9c59-46e8-bd69-cde6bd87d14a" containerName="mariadb-database-create" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.517477 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="6870aa8c-c1e3-4c3c-b1dd-c737e0a1d382" containerName="mariadb-database-create" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.517492 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="566d7edf-ec33-43e8-abd4-7e180a2fda57" containerName="keystone-db-sync" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.517511 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="49f6d2fc-6fd2-4823-befc-332a780eb6d4" containerName="ovn-config" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.517519 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="aab3ac63-a140-47f9-9fa3-113d412dd2d5" containerName="mariadb-database-create" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.518078 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2261-account-create-ndcwb" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.528190 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.570904 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-2261-account-create-ndcwb"] Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.634690 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-1a3c-account-create-kqvxn"] Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.635941 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1a3c-account-create-kqvxn" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.648414 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.675033 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqvxx\" (UniqueName: \"kubernetes.io/projected/3fe757e4-3653-4673-af9d-cf9895e28d3e-kube-api-access-cqvxx\") pod \"cinder-2261-account-create-ndcwb\" (UID: \"3fe757e4-3653-4673-af9d-cf9895e28d3e\") " pod="openstack/cinder-2261-account-create-ndcwb" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.694301 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-1a3c-account-create-kqvxn"] Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.777663 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-hx5gv"] Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.779010 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9d85d47c-hx5gv" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.780279 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqvxx\" (UniqueName: \"kubernetes.io/projected/3fe757e4-3653-4673-af9d-cf9895e28d3e-kube-api-access-cqvxx\") pod \"cinder-2261-account-create-ndcwb\" (UID: \"3fe757e4-3653-4673-af9d-cf9895e28d3e\") " pod="openstack/cinder-2261-account-create-ndcwb" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.780317 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzvpp\" (UniqueName: \"kubernetes.io/projected/75095121-6e99-4595-b06f-7440639ef8cc-kube-api-access-xzvpp\") pod \"barbican-1a3c-account-create-kqvxn\" (UID: \"75095121-6e99-4595-b06f-7440639ef8cc\") " pod="openstack/barbican-1a3c-account-create-kqvxn" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.802511 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-64vzp"] Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.803570 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-64vzp" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.813644 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.826423 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.826775 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5xp9z" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.826984 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.846985 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-hx5gv"] Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.861621 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqvxx\" (UniqueName: \"kubernetes.io/projected/3fe757e4-3653-4673-af9d-cf9895e28d3e-kube-api-access-cqvxx\") pod \"cinder-2261-account-create-ndcwb\" (UID: \"3fe757e4-3653-4673-af9d-cf9895e28d3e\") " pod="openstack/cinder-2261-account-create-ndcwb" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.864991 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-64vzp"] Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.888733 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-dns-svc\") pod \"dnsmasq-dns-5c9d85d47c-hx5gv\" (UID: \"b263770c-536f-4136-a695-a891267a9be6\") " pod="openstack/dnsmasq-dns-5c9d85d47c-hx5gv" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.888770 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzvpp\" (UniqueName: \"kubernetes.io/projected/75095121-6e99-4595-b06f-7440639ef8cc-kube-api-access-xzvpp\") pod \"barbican-1a3c-account-create-kqvxn\" (UID: \"75095121-6e99-4595-b06f-7440639ef8cc\") " pod="openstack/barbican-1a3c-account-create-kqvxn" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.888812 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9d85d47c-hx5gv\" (UID: \"b263770c-536f-4136-a695-a891267a9be6\") " pod="openstack/dnsmasq-dns-5c9d85d47c-hx5gv" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.888833 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9d85d47c-hx5gv\" (UID: \"b263770c-536f-4136-a695-a891267a9be6\") " pod="openstack/dnsmasq-dns-5c9d85d47c-hx5gv" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.888862 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-config\") pod \"dnsmasq-dns-5c9d85d47c-hx5gv\" (UID: \"b263770c-536f-4136-a695-a891267a9be6\") " pod="openstack/dnsmasq-dns-5c9d85d47c-hx5gv" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.888879 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cdp5\" (UniqueName: \"kubernetes.io/projected/b263770c-536f-4136-a695-a891267a9be6-kube-api-access-9cdp5\") pod \"dnsmasq-dns-5c9d85d47c-hx5gv\" (UID: \"b263770c-536f-4136-a695-a891267a9be6\") " pod="openstack/dnsmasq-dns-5c9d85d47c-hx5gv" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.890269 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2261-account-create-ndcwb" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.894989 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-512d-account-create-nhbjz"] Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.896090 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-512d-account-create-nhbjz" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.901694 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.921374 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-512d-account-create-nhbjz"] Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.945650 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzvpp\" (UniqueName: \"kubernetes.io/projected/75095121-6e99-4595-b06f-7440639ef8cc-kube-api-access-xzvpp\") pod \"barbican-1a3c-account-create-kqvxn\" (UID: \"75095121-6e99-4595-b06f-7440639ef8cc\") " pod="openstack/barbican-1a3c-account-create-kqvxn" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.960258 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1a3c-account-create-kqvxn" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.989905 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-dns-svc\") pod \"dnsmasq-dns-5c9d85d47c-hx5gv\" (UID: \"b263770c-536f-4136-a695-a891267a9be6\") " pod="openstack/dnsmasq-dns-5c9d85d47c-hx5gv" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.989975 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-fernet-keys\") pod \"keystone-bootstrap-64vzp\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " pod="openstack/keystone-bootstrap-64vzp" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.989996 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mztxl\" (UniqueName: \"kubernetes.io/projected/8b23db04-161e-4736-85de-a2173dfa0d4f-kube-api-access-mztxl\") pod \"keystone-bootstrap-64vzp\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " pod="openstack/keystone-bootstrap-64vzp" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.990032 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9d85d47c-hx5gv\" (UID: \"b263770c-536f-4136-a695-a891267a9be6\") " pod="openstack/dnsmasq-dns-5c9d85d47c-hx5gv" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.990053 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9d85d47c-hx5gv\" (UID: \"b263770c-536f-4136-a695-a891267a9be6\") " pod="openstack/dnsmasq-dns-5c9d85d47c-hx5gv" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.990077 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-combined-ca-bundle\") pod \"keystone-bootstrap-64vzp\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " pod="openstack/keystone-bootstrap-64vzp" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.990092 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-credential-keys\") pod \"keystone-bootstrap-64vzp\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " pod="openstack/keystone-bootstrap-64vzp" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.990111 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmfwl\" (UniqueName: \"kubernetes.io/projected/373ea549-fc1c-43eb-8af2-6feb43dd9f70-kube-api-access-fmfwl\") pod \"neutron-512d-account-create-nhbjz\" (UID: \"373ea549-fc1c-43eb-8af2-6feb43dd9f70\") " pod="openstack/neutron-512d-account-create-nhbjz" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.990134 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-config\") pod \"dnsmasq-dns-5c9d85d47c-hx5gv\" (UID: \"b263770c-536f-4136-a695-a891267a9be6\") " pod="openstack/dnsmasq-dns-5c9d85d47c-hx5gv" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.990151 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cdp5\" (UniqueName: \"kubernetes.io/projected/b263770c-536f-4136-a695-a891267a9be6-kube-api-access-9cdp5\") pod \"dnsmasq-dns-5c9d85d47c-hx5gv\" (UID: \"b263770c-536f-4136-a695-a891267a9be6\") " pod="openstack/dnsmasq-dns-5c9d85d47c-hx5gv" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.990214 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-scripts\") pod \"keystone-bootstrap-64vzp\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " pod="openstack/keystone-bootstrap-64vzp" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.990263 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-config-data\") pod \"keystone-bootstrap-64vzp\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " pod="openstack/keystone-bootstrap-64vzp" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.991094 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9d85d47c-hx5gv\" (UID: \"b263770c-536f-4136-a695-a891267a9be6\") " pod="openstack/dnsmasq-dns-5c9d85d47c-hx5gv" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.991451 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-dns-svc\") pod \"dnsmasq-dns-5c9d85d47c-hx5gv\" (UID: \"b263770c-536f-4136-a695-a891267a9be6\") " pod="openstack/dnsmasq-dns-5c9d85d47c-hx5gv" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.991711 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9d85d47c-hx5gv\" (UID: \"b263770c-536f-4136-a695-a891267a9be6\") " pod="openstack/dnsmasq-dns-5c9d85d47c-hx5gv" Oct 13 09:01:25 crc kubenswrapper[4685]: I1013 09:01:25.992271 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-config\") pod \"dnsmasq-dns-5c9d85d47c-hx5gv\" (UID: \"b263770c-536f-4136-a695-a891267a9be6\") " pod="openstack/dnsmasq-dns-5c9d85d47c-hx5gv" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.030571 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cdp5\" (UniqueName: \"kubernetes.io/projected/b263770c-536f-4136-a695-a891267a9be6-kube-api-access-9cdp5\") pod \"dnsmasq-dns-5c9d85d47c-hx5gv\" (UID: \"b263770c-536f-4136-a695-a891267a9be6\") " pod="openstack/dnsmasq-dns-5c9d85d47c-hx5gv" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.061082 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7bdb58b6d9-xrlmx"] Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.088182 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bdb58b6d9-xrlmx" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.098278 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.100022 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-config-data\") pod \"keystone-bootstrap-64vzp\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " pod="openstack/keystone-bootstrap-64vzp" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.147108 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-fernet-keys\") pod \"keystone-bootstrap-64vzp\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " pod="openstack/keystone-bootstrap-64vzp" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.150121 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mztxl\" (UniqueName: \"kubernetes.io/projected/8b23db04-161e-4736-85de-a2173dfa0d4f-kube-api-access-mztxl\") pod \"keystone-bootstrap-64vzp\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " pod="openstack/keystone-bootstrap-64vzp" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.126287 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7bdb58b6d9-xrlmx"] Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.145368 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-config-data\") pod \"keystone-bootstrap-64vzp\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " pod="openstack/keystone-bootstrap-64vzp" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.114718 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.123432 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9d85d47c-hx5gv" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.120616 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-vxdv7" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.120676 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.153355 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-combined-ca-bundle\") pod \"keystone-bootstrap-64vzp\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " pod="openstack/keystone-bootstrap-64vzp" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.153452 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-credential-keys\") pod \"keystone-bootstrap-64vzp\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " pod="openstack/keystone-bootstrap-64vzp" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.156431 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmfwl\" (UniqueName: \"kubernetes.io/projected/373ea549-fc1c-43eb-8af2-6feb43dd9f70-kube-api-access-fmfwl\") pod \"neutron-512d-account-create-nhbjz\" (UID: \"373ea549-fc1c-43eb-8af2-6feb43dd9f70\") " pod="openstack/neutron-512d-account-create-nhbjz" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.156714 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-scripts\") pod \"keystone-bootstrap-64vzp\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " pod="openstack/keystone-bootstrap-64vzp" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.158750 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-fernet-keys\") pod \"keystone-bootstrap-64vzp\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " pod="openstack/keystone-bootstrap-64vzp" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.168156 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-scripts\") pod \"keystone-bootstrap-64vzp\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " pod="openstack/keystone-bootstrap-64vzp" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.170636 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-credential-keys\") pod \"keystone-bootstrap-64vzp\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " pod="openstack/keystone-bootstrap-64vzp" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.199689 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-combined-ca-bundle\") pod \"keystone-bootstrap-64vzp\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " pod="openstack/keystone-bootstrap-64vzp" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.200625 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mztxl\" (UniqueName: \"kubernetes.io/projected/8b23db04-161e-4736-85de-a2173dfa0d4f-kube-api-access-mztxl\") pod \"keystone-bootstrap-64vzp\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " pod="openstack/keystone-bootstrap-64vzp" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.208294 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.209138 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmfwl\" (UniqueName: \"kubernetes.io/projected/373ea549-fc1c-43eb-8af2-6feb43dd9f70-kube-api-access-fmfwl\") pod \"neutron-512d-account-create-nhbjz\" (UID: \"373ea549-fc1c-43eb-8af2-6feb43dd9f70\") " pod="openstack/neutron-512d-account-create-nhbjz" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.210524 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.212846 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.212995 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.224566 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-hx5gv"] Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.250201 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.259593 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-config-data\") pod \"ceilometer-0\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.259652 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-config-data\") pod \"horizon-7bdb58b6d9-xrlmx\" (UID: \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\") " pod="openstack/horizon-7bdb58b6d9-xrlmx" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.259683 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-scripts\") pod \"ceilometer-0\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.259706 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-scripts\") pod \"horizon-7bdb58b6d9-xrlmx\" (UID: \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\") " pod="openstack/horizon-7bdb58b6d9-xrlmx" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.259742 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.259766 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-horizon-secret-key\") pod \"horizon-7bdb58b6d9-xrlmx\" (UID: \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\") " pod="openstack/horizon-7bdb58b6d9-xrlmx" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.259793 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stwmz\" (UniqueName: \"kubernetes.io/projected/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-kube-api-access-stwmz\") pod \"horizon-7bdb58b6d9-xrlmx\" (UID: \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\") " pod="openstack/horizon-7bdb58b6d9-xrlmx" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.259810 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.259851 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-run-httpd\") pod \"ceilometer-0\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.259871 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4fhq\" (UniqueName: \"kubernetes.io/projected/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-kube-api-access-c4fhq\") pod \"ceilometer-0\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.259887 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-log-httpd\") pod \"ceilometer-0\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.259908 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-logs\") pod \"horizon-7bdb58b6d9-xrlmx\" (UID: \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\") " pod="openstack/horizon-7bdb58b6d9-xrlmx" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.273362 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b868669f-ds2mk"] Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.274790 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-ds2mk" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.282340 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.294191 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-4phvd"] Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.297195 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4phvd" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.309346 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-512d-account-create-nhbjz" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.309830 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.309940 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-cfczb" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.315797 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.349526 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-4phvd"] Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.367934 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.367981 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-horizon-secret-key\") pod \"horizon-7bdb58b6d9-xrlmx\" (UID: \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\") " pod="openstack/horizon-7bdb58b6d9-xrlmx" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.368023 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5afcbfe-3193-426b-97fe-4e43f4a483d8-scripts\") pod \"placement-db-sync-4phvd\" (UID: \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\") " pod="openstack/placement-db-sync-4phvd" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.368042 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stwmz\" (UniqueName: \"kubernetes.io/projected/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-kube-api-access-stwmz\") pod \"horizon-7bdb58b6d9-xrlmx\" (UID: \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\") " pod="openstack/horizon-7bdb58b6d9-xrlmx" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.368062 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.368099 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-config\") pod \"dnsmasq-dns-5b868669f-ds2mk\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " pod="openstack/dnsmasq-dns-5b868669f-ds2mk" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.368119 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-ds2mk\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " pod="openstack/dnsmasq-dns-5b868669f-ds2mk" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.368364 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-run-httpd\") pod \"ceilometer-0\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.368391 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4fhq\" (UniqueName: \"kubernetes.io/projected/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-kube-api-access-c4fhq\") pod \"ceilometer-0\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.368408 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-log-httpd\") pod \"ceilometer-0\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.368443 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-ds2mk\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " pod="openstack/dnsmasq-dns-5b868669f-ds2mk" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.368460 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-ds2mk\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " pod="openstack/dnsmasq-dns-5b868669f-ds2mk" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.368475 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-logs\") pod \"horizon-7bdb58b6d9-xrlmx\" (UID: \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\") " pod="openstack/horizon-7bdb58b6d9-xrlmx" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.368491 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5afcbfe-3193-426b-97fe-4e43f4a483d8-config-data\") pod \"placement-db-sync-4phvd\" (UID: \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\") " pod="openstack/placement-db-sync-4phvd" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.368528 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcb7p\" (UniqueName: \"kubernetes.io/projected/c5afcbfe-3193-426b-97fe-4e43f4a483d8-kube-api-access-kcb7p\") pod \"placement-db-sync-4phvd\" (UID: \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\") " pod="openstack/placement-db-sync-4phvd" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.368554 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-config-data\") pod \"ceilometer-0\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.368591 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-config-data\") pod \"horizon-7bdb58b6d9-xrlmx\" (UID: \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\") " pod="openstack/horizon-7bdb58b6d9-xrlmx" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.368612 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-scripts\") pod \"ceilometer-0\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.368632 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-scripts\") pod \"horizon-7bdb58b6d9-xrlmx\" (UID: \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\") " pod="openstack/horizon-7bdb58b6d9-xrlmx" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.368670 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbc6f\" (UniqueName: \"kubernetes.io/projected/b2f96ed0-e090-432c-b72e-61de91c7f11e-kube-api-access-tbc6f\") pod \"dnsmasq-dns-5b868669f-ds2mk\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " pod="openstack/dnsmasq-dns-5b868669f-ds2mk" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.368688 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5afcbfe-3193-426b-97fe-4e43f4a483d8-combined-ca-bundle\") pod \"placement-db-sync-4phvd\" (UID: \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\") " pod="openstack/placement-db-sync-4phvd" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.368702 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5afcbfe-3193-426b-97fe-4e43f4a483d8-logs\") pod \"placement-db-sync-4phvd\" (UID: \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\") " pod="openstack/placement-db-sync-4phvd" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.368720 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-dns-svc\") pod \"dnsmasq-dns-5b868669f-ds2mk\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " pod="openstack/dnsmasq-dns-5b868669f-ds2mk" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.373375 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-run-httpd\") pod \"ceilometer-0\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.373866 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-log-httpd\") pod \"ceilometer-0\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.374252 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-logs\") pod \"horizon-7bdb58b6d9-xrlmx\" (UID: \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\") " pod="openstack/horizon-7bdb58b6d9-xrlmx" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.376274 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-scripts\") pod \"horizon-7bdb58b6d9-xrlmx\" (UID: \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\") " pod="openstack/horizon-7bdb58b6d9-xrlmx" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.382038 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-ds2mk"] Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.382773 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-scripts\") pod \"ceilometer-0\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.383973 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-config-data\") pod \"horizon-7bdb58b6d9-xrlmx\" (UID: \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\") " pod="openstack/horizon-7bdb58b6d9-xrlmx" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.385755 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.391055 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.394976 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-horizon-secret-key\") pod \"horizon-7bdb58b6d9-xrlmx\" (UID: \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\") " pod="openstack/horizon-7bdb58b6d9-xrlmx" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.404938 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4fhq\" (UniqueName: \"kubernetes.io/projected/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-kube-api-access-c4fhq\") pod \"ceilometer-0\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.417233 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-config-data\") pod \"ceilometer-0\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.422005 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stwmz\" (UniqueName: \"kubernetes.io/projected/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-kube-api-access-stwmz\") pod \"horizon-7bdb58b6d9-xrlmx\" (UID: \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\") " pod="openstack/horizon-7bdb58b6d9-xrlmx" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.446070 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bdb58b6d9-xrlmx" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.455978 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-ds2mk"] Oct 13 09:01:26 crc kubenswrapper[4685]: E1013 09:01:26.456549 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-tbc6f ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-5b868669f-ds2mk" podUID="b2f96ed0-e090-432c-b72e-61de91c7f11e" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.470776 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5afcbfe-3193-426b-97fe-4e43f4a483d8-scripts\") pod \"placement-db-sync-4phvd\" (UID: \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\") " pod="openstack/placement-db-sync-4phvd" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.470839 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-config\") pod \"dnsmasq-dns-5b868669f-ds2mk\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " pod="openstack/dnsmasq-dns-5b868669f-ds2mk" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.470857 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-ds2mk\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " pod="openstack/dnsmasq-dns-5b868669f-ds2mk" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.470993 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-ds2mk\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " pod="openstack/dnsmasq-dns-5b868669f-ds2mk" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.471011 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-ds2mk\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " pod="openstack/dnsmasq-dns-5b868669f-ds2mk" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.471026 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5afcbfe-3193-426b-97fe-4e43f4a483d8-config-data\") pod \"placement-db-sync-4phvd\" (UID: \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\") " pod="openstack/placement-db-sync-4phvd" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.471049 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcb7p\" (UniqueName: \"kubernetes.io/projected/c5afcbfe-3193-426b-97fe-4e43f4a483d8-kube-api-access-kcb7p\") pod \"placement-db-sync-4phvd\" (UID: \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\") " pod="openstack/placement-db-sync-4phvd" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.471100 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbc6f\" (UniqueName: \"kubernetes.io/projected/b2f96ed0-e090-432c-b72e-61de91c7f11e-kube-api-access-tbc6f\") pod \"dnsmasq-dns-5b868669f-ds2mk\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " pod="openstack/dnsmasq-dns-5b868669f-ds2mk" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.471121 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5afcbfe-3193-426b-97fe-4e43f4a483d8-logs\") pod \"placement-db-sync-4phvd\" (UID: \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\") " pod="openstack/placement-db-sync-4phvd" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.471137 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5afcbfe-3193-426b-97fe-4e43f4a483d8-combined-ca-bundle\") pod \"placement-db-sync-4phvd\" (UID: \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\") " pod="openstack/placement-db-sync-4phvd" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.471160 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-dns-svc\") pod \"dnsmasq-dns-5b868669f-ds2mk\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " pod="openstack/dnsmasq-dns-5b868669f-ds2mk" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.471960 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-dns-svc\") pod \"dnsmasq-dns-5b868669f-ds2mk\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " pod="openstack/dnsmasq-dns-5b868669f-ds2mk" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.474146 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-ds2mk\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " pod="openstack/dnsmasq-dns-5b868669f-ds2mk" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.474726 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-config\") pod \"dnsmasq-dns-5b868669f-ds2mk\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " pod="openstack/dnsmasq-dns-5b868669f-ds2mk" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.475296 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-ds2mk\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " pod="openstack/dnsmasq-dns-5b868669f-ds2mk" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.475886 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-ds2mk\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " pod="openstack/dnsmasq-dns-5b868669f-ds2mk" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.476549 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5afcbfe-3193-426b-97fe-4e43f4a483d8-logs\") pod \"placement-db-sync-4phvd\" (UID: \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\") " pod="openstack/placement-db-sync-4phvd" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.481045 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5afcbfe-3193-426b-97fe-4e43f4a483d8-scripts\") pod \"placement-db-sync-4phvd\" (UID: \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\") " pod="openstack/placement-db-sync-4phvd" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.490241 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-64vzp" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.495977 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-ll5t2"] Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.496247 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5afcbfe-3193-426b-97fe-4e43f4a483d8-combined-ca-bundle\") pod \"placement-db-sync-4phvd\" (UID: \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\") " pod="openstack/placement-db-sync-4phvd" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.498164 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5afcbfe-3193-426b-97fe-4e43f4a483d8-config-data\") pod \"placement-db-sync-4phvd\" (UID: \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\") " pod="openstack/placement-db-sync-4phvd" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.508418 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcb7p\" (UniqueName: \"kubernetes.io/projected/c5afcbfe-3193-426b-97fe-4e43f4a483d8-kube-api-access-kcb7p\") pod \"placement-db-sync-4phvd\" (UID: \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\") " pod="openstack/placement-db-sync-4phvd" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.509640 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.513361 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-ll5t2"] Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.520425 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-695cd469df-pgg2b"] Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.529299 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-695cd469df-pgg2b"] Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.531213 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jvxxs" event={"ID":"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251","Type":"ContainerDied","Data":"3f7fb51165e65e09f0de8b115d50f44b4dca84584b4d149722bbd704aa2a4718"} Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.529515 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-695cd469df-pgg2b" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.529811 4685 generic.go:334] "Generic (PLEG): container finished" podID="b5a0947d-3515-4e7c-a69f-c7e4d3e9f251" containerID="3f7fb51165e65e09f0de8b115d50f44b4dca84584b4d149722bbd704aa2a4718" exitCode=0 Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.540640 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.549816 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbc6f\" (UniqueName: \"kubernetes.io/projected/b2f96ed0-e090-432c-b72e-61de91c7f11e-kube-api-access-tbc6f\") pod \"dnsmasq-dns-5b868669f-ds2mk\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " pod="openstack/dnsmasq-dns-5b868669f-ds2mk" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.573636 4685 generic.go:334] "Generic (PLEG): container finished" podID="b001d17a-1aea-44ba-86c5-ba6b312156c1" containerID="1b0c736e43589c77b3b7ea005ec8f55bd09664ac09cc3b4b417ff1bad387941b" exitCode=1 Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.573854 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-ll5t2\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.573893 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-ll5t2\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.573967 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48006700-a390-415e-97c6-c083512b1fb7-scripts\") pod \"horizon-695cd469df-pgg2b\" (UID: \"48006700-a390-415e-97c6-c083512b1fb7\") " pod="openstack/horizon-695cd469df-pgg2b" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.573999 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48006700-a390-415e-97c6-c083512b1fb7-logs\") pod \"horizon-695cd469df-pgg2b\" (UID: \"48006700-a390-415e-97c6-c083512b1fb7\") " pod="openstack/horizon-695cd469df-pgg2b" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.574015 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-dns-svc\") pod \"dnsmasq-dns-cf78879c9-ll5t2\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.574049 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-config\") pod \"dnsmasq-dns-cf78879c9-ll5t2\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.574077 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwnf8\" (UniqueName: \"kubernetes.io/projected/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-kube-api-access-hwnf8\") pod \"dnsmasq-dns-cf78879c9-ll5t2\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.574098 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/48006700-a390-415e-97c6-c083512b1fb7-horizon-secret-key\") pod \"horizon-695cd469df-pgg2b\" (UID: \"48006700-a390-415e-97c6-c083512b1fb7\") " pod="openstack/horizon-695cd469df-pgg2b" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.574128 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/48006700-a390-415e-97c6-c083512b1fb7-config-data\") pod \"horizon-695cd469df-pgg2b\" (UID: \"48006700-a390-415e-97c6-c083512b1fb7\") " pod="openstack/horizon-695cd469df-pgg2b" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.574146 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snxdd\" (UniqueName: \"kubernetes.io/projected/48006700-a390-415e-97c6-c083512b1fb7-kube-api-access-snxdd\") pod \"horizon-695cd469df-pgg2b\" (UID: \"48006700-a390-415e-97c6-c083512b1fb7\") " pod="openstack/horizon-695cd469df-pgg2b" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.574169 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-ll5t2\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.574771 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" event={"ID":"b001d17a-1aea-44ba-86c5-ba6b312156c1","Type":"ContainerDied","Data":"1b0c736e43589c77b3b7ea005ec8f55bd09664ac09cc3b4b417ff1bad387941b"} Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.575061 4685 scope.go:117] "RemoveContainer" containerID="1b0c736e43589c77b3b7ea005ec8f55bd09664ac09cc3b4b417ff1bad387941b" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.644714 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4phvd" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.686123 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/48006700-a390-415e-97c6-c083512b1fb7-config-data\") pod \"horizon-695cd469df-pgg2b\" (UID: \"48006700-a390-415e-97c6-c083512b1fb7\") " pod="openstack/horizon-695cd469df-pgg2b" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.686163 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snxdd\" (UniqueName: \"kubernetes.io/projected/48006700-a390-415e-97c6-c083512b1fb7-kube-api-access-snxdd\") pod \"horizon-695cd469df-pgg2b\" (UID: \"48006700-a390-415e-97c6-c083512b1fb7\") " pod="openstack/horizon-695cd469df-pgg2b" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.686192 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-ll5t2\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.686244 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-ll5t2\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.686262 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-ll5t2\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.686311 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48006700-a390-415e-97c6-c083512b1fb7-scripts\") pod \"horizon-695cd469df-pgg2b\" (UID: \"48006700-a390-415e-97c6-c083512b1fb7\") " pod="openstack/horizon-695cd469df-pgg2b" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.686348 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48006700-a390-415e-97c6-c083512b1fb7-logs\") pod \"horizon-695cd469df-pgg2b\" (UID: \"48006700-a390-415e-97c6-c083512b1fb7\") " pod="openstack/horizon-695cd469df-pgg2b" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.686363 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-dns-svc\") pod \"dnsmasq-dns-cf78879c9-ll5t2\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.686396 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-config\") pod \"dnsmasq-dns-cf78879c9-ll5t2\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.686422 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwnf8\" (UniqueName: \"kubernetes.io/projected/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-kube-api-access-hwnf8\") pod \"dnsmasq-dns-cf78879c9-ll5t2\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.686441 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/48006700-a390-415e-97c6-c083512b1fb7-horizon-secret-key\") pod \"horizon-695cd469df-pgg2b\" (UID: \"48006700-a390-415e-97c6-c083512b1fb7\") " pod="openstack/horizon-695cd469df-pgg2b" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.687233 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48006700-a390-415e-97c6-c083512b1fb7-logs\") pod \"horizon-695cd469df-pgg2b\" (UID: \"48006700-a390-415e-97c6-c083512b1fb7\") " pod="openstack/horizon-695cd469df-pgg2b" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.687715 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-ll5t2\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.687787 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48006700-a390-415e-97c6-c083512b1fb7-scripts\") pod \"horizon-695cd469df-pgg2b\" (UID: \"48006700-a390-415e-97c6-c083512b1fb7\") " pod="openstack/horizon-695cd469df-pgg2b" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.688985 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-ll5t2\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.689001 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/48006700-a390-415e-97c6-c083512b1fb7-config-data\") pod \"horizon-695cd469df-pgg2b\" (UID: \"48006700-a390-415e-97c6-c083512b1fb7\") " pod="openstack/horizon-695cd469df-pgg2b" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.689882 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-config\") pod \"dnsmasq-dns-cf78879c9-ll5t2\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.690238 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/48006700-a390-415e-97c6-c083512b1fb7-horizon-secret-key\") pod \"horizon-695cd469df-pgg2b\" (UID: \"48006700-a390-415e-97c6-c083512b1fb7\") " pod="openstack/horizon-695cd469df-pgg2b" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.690538 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-dns-svc\") pod \"dnsmasq-dns-cf78879c9-ll5t2\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.699439 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-ll5t2\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.767115 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwnf8\" (UniqueName: \"kubernetes.io/projected/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-kube-api-access-hwnf8\") pod \"dnsmasq-dns-cf78879c9-ll5t2\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:26 crc kubenswrapper[4685]: I1013 09:01:26.814553 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snxdd\" (UniqueName: \"kubernetes.io/projected/48006700-a390-415e-97c6-c083512b1fb7-kube-api-access-snxdd\") pod \"horizon-695cd469df-pgg2b\" (UID: \"48006700-a390-415e-97c6-c083512b1fb7\") " pod="openstack/horizon-695cd469df-pgg2b" Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.008182 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-2261-account-create-ndcwb"] Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.055453 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.078142 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-695cd469df-pgg2b" Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.150175 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-1a3c-account-create-kqvxn"] Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.621651 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" event={"ID":"b001d17a-1aea-44ba-86c5-ba6b312156c1","Type":"ContainerStarted","Data":"c44b3a507602babf650e62c8b2c5a5ac1d0f13d11239f82857414365cf156ea7"} Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.623576 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.653147 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2261-account-create-ndcwb" event={"ID":"3fe757e4-3653-4673-af9d-cf9895e28d3e","Type":"ContainerStarted","Data":"9543a6b457576086f09a33ada106d2ef6ec6ee0377517be3c9e7e084fe7818b1"} Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.660228 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1a3c-account-create-kqvxn" event={"ID":"75095121-6e99-4595-b06f-7440639ef8cc","Type":"ContainerStarted","Data":"06aa14aea746bcb1472bf6181bd00dfeeccc2b52736d1a22d5a834baaf85bc37"} Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.660288 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-ds2mk" Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.687879 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-ds2mk" Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.737322 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-dns-svc\") pod \"b2f96ed0-e090-432c-b72e-61de91c7f11e\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.737505 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbc6f\" (UniqueName: \"kubernetes.io/projected/b2f96ed0-e090-432c-b72e-61de91c7f11e-kube-api-access-tbc6f\") pod \"b2f96ed0-e090-432c-b72e-61de91c7f11e\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.737530 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-config\") pod \"b2f96ed0-e090-432c-b72e-61de91c7f11e\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.737550 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-ovsdbserver-nb\") pod \"b2f96ed0-e090-432c-b72e-61de91c7f11e\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.737602 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-dns-swift-storage-0\") pod \"b2f96ed0-e090-432c-b72e-61de91c7f11e\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.737669 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-ovsdbserver-sb\") pod \"b2f96ed0-e090-432c-b72e-61de91c7f11e\" (UID: \"b2f96ed0-e090-432c-b72e-61de91c7f11e\") " Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.738700 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-config" (OuterVolumeSpecName: "config") pod "b2f96ed0-e090-432c-b72e-61de91c7f11e" (UID: "b2f96ed0-e090-432c-b72e-61de91c7f11e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.739178 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b2f96ed0-e090-432c-b72e-61de91c7f11e" (UID: "b2f96ed0-e090-432c-b72e-61de91c7f11e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.739541 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b2f96ed0-e090-432c-b72e-61de91c7f11e" (UID: "b2f96ed0-e090-432c-b72e-61de91c7f11e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.740032 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b2f96ed0-e090-432c-b72e-61de91c7f11e" (UID: "b2f96ed0-e090-432c-b72e-61de91c7f11e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.740509 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b2f96ed0-e090-432c-b72e-61de91c7f11e" (UID: "b2f96ed0-e090-432c-b72e-61de91c7f11e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.760010 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2f96ed0-e090-432c-b72e-61de91c7f11e-kube-api-access-tbc6f" (OuterVolumeSpecName: "kube-api-access-tbc6f") pod "b2f96ed0-e090-432c-b72e-61de91c7f11e" (UID: "b2f96ed0-e090-432c-b72e-61de91c7f11e"). InnerVolumeSpecName "kube-api-access-tbc6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.839457 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.839498 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbc6f\" (UniqueName: \"kubernetes.io/projected/b2f96ed0-e090-432c-b72e-61de91c7f11e-kube-api-access-tbc6f\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.839509 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.839519 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-config\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.839529 4685 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.839537 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2f96ed0-e090-432c-b72e-61de91c7f11e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:27 crc kubenswrapper[4685]: I1013 09:01:27.970754 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-512d-account-create-nhbjz"] Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.492481 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-64vzp"] Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.550239 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-ll5t2"] Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.555853 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-695cd469df-pgg2b"] Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.583414 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-hx5gv"] Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.639547 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.659951 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-4phvd"] Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.672853 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7bdb58b6d9-xrlmx"] Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.709082 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" event={"ID":"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0","Type":"ContainerStarted","Data":"d9cdce5d38835674053bdee52c49f671b8fb95326570247e88174c832833f6e1"} Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.735581 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-512d-account-create-nhbjz" event={"ID":"373ea549-fc1c-43eb-8af2-6feb43dd9f70","Type":"ContainerStarted","Data":"54524e8771985cb5390931a8625210d02bfffcc3b1af2e7fdb3cb825339da6b6"} Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.735704 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-512d-account-create-nhbjz" event={"ID":"373ea549-fc1c-43eb-8af2-6feb43dd9f70","Type":"ContainerStarted","Data":"b4737dffc7c4f0bf25cff55e746bf42a9a3a463c90decae9dc0de13117662053"} Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.743066 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-695cd469df-pgg2b" event={"ID":"48006700-a390-415e-97c6-c083512b1fb7","Type":"ContainerStarted","Data":"c581ae18eaadd387e55dadb02db33cfb45a858754ce32a46ebacb32c2f93901b"} Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.748093 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-64vzp" event={"ID":"8b23db04-161e-4736-85de-a2173dfa0d4f","Type":"ContainerStarted","Data":"3155b9a33d1ca7af94ce2d6c23557d9403654ec051d08d9e85e3b8b5beb72de6"} Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.752255 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ffc408c-86d3-4e93-9d49-6ef39cb7d959","Type":"ContainerStarted","Data":"0306714f5cc052aeb656ce1d1a1366c895a35ae0f5259e364656562c360fa75e"} Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.763394 4685 generic.go:334] "Generic (PLEG): container finished" podID="75095121-6e99-4595-b06f-7440639ef8cc" containerID="fc673b5b34bc03076fb56efe1827e2a6d8cc8e9e874ad8f9a0c7d2900481399f" exitCode=0 Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.763447 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1a3c-account-create-kqvxn" event={"ID":"75095121-6e99-4595-b06f-7440639ef8cc","Type":"ContainerDied","Data":"fc673b5b34bc03076fb56efe1827e2a6d8cc8e9e874ad8f9a0c7d2900481399f"} Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.783966 4685 generic.go:334] "Generic (PLEG): container finished" podID="3fe757e4-3653-4673-af9d-cf9895e28d3e" containerID="5977da2ef9ebca4c1e0e11708be68c1d3bf8839636a1b5ffde638454afd0c394" exitCode=0 Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.784070 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2261-account-create-ndcwb" event={"ID":"3fe757e4-3653-4673-af9d-cf9895e28d3e","Type":"ContainerDied","Data":"5977da2ef9ebca4c1e0e11708be68c1d3bf8839636a1b5ffde638454afd0c394"} Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.795904 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jvxxs" event={"ID":"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251","Type":"ContainerDied","Data":"efb67fa5579071bb13f37055a2e4accfee710a1916b236c22b47a52d88b4f77b"} Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.795962 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="efb67fa5579071bb13f37055a2e4accfee710a1916b236c22b47a52d88b4f77b" Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.814283 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9d85d47c-hx5gv" event={"ID":"b263770c-536f-4136-a695-a891267a9be6","Type":"ContainerStarted","Data":"8ea9df880da97e34cecdc4b67401db1e09d1e015db51043180fe09b549590e24"} Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.814337 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-ds2mk" Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.852619 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jvxxs" Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.864848 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-combined-ca-bundle\") pod \"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251\" (UID: \"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251\") " Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.865044 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-db-sync-config-data\") pod \"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251\" (UID: \"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251\") " Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.865127 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-config-data\") pod \"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251\" (UID: \"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251\") " Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.865204 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cthk7\" (UniqueName: \"kubernetes.io/projected/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-kube-api-access-cthk7\") pod \"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251\" (UID: \"b5a0947d-3515-4e7c-a69f-c7e4d3e9f251\") " Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.903090 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-kube-api-access-cthk7" (OuterVolumeSpecName: "kube-api-access-cthk7") pod "b5a0947d-3515-4e7c-a69f-c7e4d3e9f251" (UID: "b5a0947d-3515-4e7c-a69f-c7e4d3e9f251"). InnerVolumeSpecName "kube-api-access-cthk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.903513 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b5a0947d-3515-4e7c-a69f-c7e4d3e9f251" (UID: "b5a0947d-3515-4e7c-a69f-c7e4d3e9f251"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.969712 4685 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.969977 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cthk7\" (UniqueName: \"kubernetes.io/projected/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-kube-api-access-cthk7\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.977872 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-ds2mk"] Oct 13 09:01:28 crc kubenswrapper[4685]: I1013 09:01:28.988447 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-ds2mk"] Oct 13 09:01:29 crc kubenswrapper[4685]: I1013 09:01:29.061170 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b5a0947d-3515-4e7c-a69f-c7e4d3e9f251" (UID: "b5a0947d-3515-4e7c-a69f-c7e4d3e9f251"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:29 crc kubenswrapper[4685]: I1013 09:01:29.073708 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:29 crc kubenswrapper[4685]: I1013 09:01:29.085579 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-config-data" (OuterVolumeSpecName: "config-data") pod "b5a0947d-3515-4e7c-a69f-c7e4d3e9f251" (UID: "b5a0947d-3515-4e7c-a69f-c7e4d3e9f251"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:29 crc kubenswrapper[4685]: I1013 09:01:29.175633 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:29 crc kubenswrapper[4685]: I1013 09:01:29.514857 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2f96ed0-e090-432c-b72e-61de91c7f11e" path="/var/lib/kubelet/pods/b2f96ed0-e090-432c-b72e-61de91c7f11e/volumes" Oct 13 09:01:29 crc kubenswrapper[4685]: I1013 09:01:29.827412 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bdb58b6d9-xrlmx" event={"ID":"552e9180-01ac-4cbc-9060-2fa6e6e18e9f","Type":"ContainerStarted","Data":"eeb7340908b4e870a0499fad831954f05d8f9f158c9a7687cfb1ae1cc6cc898f"} Oct 13 09:01:29 crc kubenswrapper[4685]: I1013 09:01:29.829744 4685 generic.go:334] "Generic (PLEG): container finished" podID="373ea549-fc1c-43eb-8af2-6feb43dd9f70" containerID="54524e8771985cb5390931a8625210d02bfffcc3b1af2e7fdb3cb825339da6b6" exitCode=0 Oct 13 09:01:29 crc kubenswrapper[4685]: I1013 09:01:29.829831 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-512d-account-create-nhbjz" event={"ID":"373ea549-fc1c-43eb-8af2-6feb43dd9f70","Type":"ContainerDied","Data":"54524e8771985cb5390931a8625210d02bfffcc3b1af2e7fdb3cb825339da6b6"} Oct 13 09:01:29 crc kubenswrapper[4685]: I1013 09:01:29.835493 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-64vzp" event={"ID":"8b23db04-161e-4736-85de-a2173dfa0d4f","Type":"ContainerStarted","Data":"09eb66b97da1f79efa9078a0e2afd34299176faadf74af83bf9b0b03d570003f"} Oct 13 09:01:29 crc kubenswrapper[4685]: I1013 09:01:29.838616 4685 generic.go:334] "Generic (PLEG): container finished" podID="b263770c-536f-4136-a695-a891267a9be6" containerID="ee5678ba0a208466b145c2bbc93068e9a0828d4eb192866c67b7e5897fc461a0" exitCode=0 Oct 13 09:01:29 crc kubenswrapper[4685]: I1013 09:01:29.838716 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9d85d47c-hx5gv" event={"ID":"b263770c-536f-4136-a695-a891267a9be6","Type":"ContainerDied","Data":"ee5678ba0a208466b145c2bbc93068e9a0828d4eb192866c67b7e5897fc461a0"} Oct 13 09:01:29 crc kubenswrapper[4685]: I1013 09:01:29.850201 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4phvd" event={"ID":"c5afcbfe-3193-426b-97fe-4e43f4a483d8","Type":"ContainerStarted","Data":"e2311f145362707a4f3ea678c410be719e938553a3ce91d16679b111b8281f36"} Oct 13 09:01:29 crc kubenswrapper[4685]: I1013 09:01:29.853195 4685 generic.go:334] "Generic (PLEG): container finished" podID="c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0" containerID="3aa2c3741b4424b20ec1e164a1ce790ea7358079b0d22bc17e17e025af91175f" exitCode=0 Oct 13 09:01:29 crc kubenswrapper[4685]: I1013 09:01:29.853802 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" event={"ID":"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0","Type":"ContainerDied","Data":"3aa2c3741b4424b20ec1e164a1ce790ea7358079b0d22bc17e17e025af91175f"} Oct 13 09:01:29 crc kubenswrapper[4685]: I1013 09:01:29.853866 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jvxxs" Oct 13 09:01:29 crc kubenswrapper[4685]: I1013 09:01:29.870990 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-64vzp" podStartSLOduration=4.870970835 podStartE2EDuration="4.870970835s" podCreationTimestamp="2025-10-13 09:01:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:01:29.862907065 +0000 UTC m=+1015.010782826" watchObservedRunningTime="2025-10-13 09:01:29.870970835 +0000 UTC m=+1015.018846606" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.285774 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-512d-account-create-nhbjz" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.315090 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmfwl\" (UniqueName: \"kubernetes.io/projected/373ea549-fc1c-43eb-8af2-6feb43dd9f70-kube-api-access-fmfwl\") pod \"373ea549-fc1c-43eb-8af2-6feb43dd9f70\" (UID: \"373ea549-fc1c-43eb-8af2-6feb43dd9f70\") " Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.323562 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/373ea549-fc1c-43eb-8af2-6feb43dd9f70-kube-api-access-fmfwl" (OuterVolumeSpecName: "kube-api-access-fmfwl") pod "373ea549-fc1c-43eb-8af2-6feb43dd9f70" (UID: "373ea549-fc1c-43eb-8af2-6feb43dd9f70"). InnerVolumeSpecName "kube-api-access-fmfwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.423685 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmfwl\" (UniqueName: \"kubernetes.io/projected/373ea549-fc1c-43eb-8af2-6feb43dd9f70-kube-api-access-fmfwl\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.572081 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-ll5t2"] Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.610405 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-8hg5b"] Oct 13 09:01:30 crc kubenswrapper[4685]: E1013 09:01:30.610739 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="373ea549-fc1c-43eb-8af2-6feb43dd9f70" containerName="mariadb-account-create" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.610756 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="373ea549-fc1c-43eb-8af2-6feb43dd9f70" containerName="mariadb-account-create" Oct 13 09:01:30 crc kubenswrapper[4685]: E1013 09:01:30.610775 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5a0947d-3515-4e7c-a69f-c7e4d3e9f251" containerName="glance-db-sync" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.610781 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5a0947d-3515-4e7c-a69f-c7e4d3e9f251" containerName="glance-db-sync" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.610965 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="373ea549-fc1c-43eb-8af2-6feb43dd9f70" containerName="mariadb-account-create" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.610989 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5a0947d-3515-4e7c-a69f-c7e4d3e9f251" containerName="glance-db-sync" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.619925 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.653361 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-8hg5b"] Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.658611 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9d85d47c-hx5gv" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.726490 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-dns-svc\") pod \"b263770c-536f-4136-a695-a891267a9be6\" (UID: \"b263770c-536f-4136-a695-a891267a9be6\") " Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.726601 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cdp5\" (UniqueName: \"kubernetes.io/projected/b263770c-536f-4136-a695-a891267a9be6-kube-api-access-9cdp5\") pod \"b263770c-536f-4136-a695-a891267a9be6\" (UID: \"b263770c-536f-4136-a695-a891267a9be6\") " Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.726650 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-ovsdbserver-sb\") pod \"b263770c-536f-4136-a695-a891267a9be6\" (UID: \"b263770c-536f-4136-a695-a891267a9be6\") " Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.726689 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-config\") pod \"b263770c-536f-4136-a695-a891267a9be6\" (UID: \"b263770c-536f-4136-a695-a891267a9be6\") " Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.726750 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-ovsdbserver-nb\") pod \"b263770c-536f-4136-a695-a891267a9be6\" (UID: \"b263770c-536f-4136-a695-a891267a9be6\") " Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.727012 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-8hg5b\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.727048 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-8hg5b\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.727074 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-config\") pod \"dnsmasq-dns-56df8fb6b7-8hg5b\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.727090 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-8hg5b\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.727158 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c5x2\" (UniqueName: \"kubernetes.io/projected/778d19c2-2541-40d9-a00c-efb46a746a32-kube-api-access-7c5x2\") pod \"dnsmasq-dns-56df8fb6b7-8hg5b\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.727212 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-8hg5b\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.763803 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b263770c-536f-4136-a695-a891267a9be6-kube-api-access-9cdp5" (OuterVolumeSpecName: "kube-api-access-9cdp5") pod "b263770c-536f-4136-a695-a891267a9be6" (UID: "b263770c-536f-4136-a695-a891267a9be6"). InnerVolumeSpecName "kube-api-access-9cdp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.824582 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b263770c-536f-4136-a695-a891267a9be6" (UID: "b263770c-536f-4136-a695-a891267a9be6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.833321 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-8hg5b\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.833392 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-8hg5b\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.833429 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-config\") pod \"dnsmasq-dns-56df8fb6b7-8hg5b\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.833447 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-8hg5b\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.833562 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c5x2\" (UniqueName: \"kubernetes.io/projected/778d19c2-2541-40d9-a00c-efb46a746a32-kube-api-access-7c5x2\") pod \"dnsmasq-dns-56df8fb6b7-8hg5b\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.833652 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-8hg5b\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.833740 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cdp5\" (UniqueName: \"kubernetes.io/projected/b263770c-536f-4136-a695-a891267a9be6-kube-api-access-9cdp5\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.833758 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.834604 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-8hg5b\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.835166 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-8hg5b\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.836125 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-8hg5b\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.840604 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-8hg5b\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.844156 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-config\") pod \"dnsmasq-dns-56df8fb6b7-8hg5b\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.862307 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c5x2\" (UniqueName: \"kubernetes.io/projected/778d19c2-2541-40d9-a00c-efb46a746a32-kube-api-access-7c5x2\") pod \"dnsmasq-dns-56df8fb6b7-8hg5b\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.869935 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b263770c-536f-4136-a695-a891267a9be6" (UID: "b263770c-536f-4136-a695-a891267a9be6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.882514 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b263770c-536f-4136-a695-a891267a9be6" (UID: "b263770c-536f-4136-a695-a891267a9be6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.904368 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-config" (OuterVolumeSpecName: "config") pod "b263770c-536f-4136-a695-a891267a9be6" (UID: "b263770c-536f-4136-a695-a891267a9be6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.921290 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2261-account-create-ndcwb" event={"ID":"3fe757e4-3653-4673-af9d-cf9895e28d3e","Type":"ContainerDied","Data":"9543a6b457576086f09a33ada106d2ef6ec6ee0377517be3c9e7e084fe7818b1"} Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.921610 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9543a6b457576086f09a33ada106d2ef6ec6ee0377517be3c9e7e084fe7818b1" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.935367 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.939757 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-config\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.939842 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b263770c-536f-4136-a695-a891267a9be6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.935511 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-512d-account-create-nhbjz" event={"ID":"373ea549-fc1c-43eb-8af2-6feb43dd9f70","Type":"ContainerDied","Data":"b4737dffc7c4f0bf25cff55e746bf42a9a3a463c90decae9dc0de13117662053"} Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.939978 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4737dffc7c4f0bf25cff55e746bf42a9a3a463c90decae9dc0de13117662053" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.935645 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-512d-account-create-nhbjz" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.947332 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9d85d47c-hx5gv" event={"ID":"b263770c-536f-4136-a695-a891267a9be6","Type":"ContainerDied","Data":"8ea9df880da97e34cecdc4b67401db1e09d1e015db51043180fe09b549590e24"} Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.947393 4685 scope.go:117] "RemoveContainer" containerID="ee5678ba0a208466b145c2bbc93068e9a0828d4eb192866c67b7e5897fc461a0" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.947561 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1a3c-account-create-kqvxn" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.947938 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9d85d47c-hx5gv" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.951008 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2261-account-create-ndcwb" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.952154 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1a3c-account-create-kqvxn" event={"ID":"75095121-6e99-4595-b06f-7440639ef8cc","Type":"ContainerDied","Data":"06aa14aea746bcb1472bf6181bd00dfeeccc2b52736d1a22d5a834baaf85bc37"} Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.952195 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06aa14aea746bcb1472bf6181bd00dfeeccc2b52736d1a22d5a834baaf85bc37" Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.971601 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" podUID="c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0" containerName="dnsmasq-dns" containerID="cri-o://afa5be8b3582675c84d7086fd92fad8ff4320fa17cfa09cdae4ea3b950b26806" gracePeriod=10 Oct 13 09:01:30 crc kubenswrapper[4685]: I1013 09:01:30.971696 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.011365 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.014427 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" podStartSLOduration=5.014405886 podStartE2EDuration="5.014405886s" podCreationTimestamp="2025-10-13 09:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:01:31.010830789 +0000 UTC m=+1016.158706570" watchObservedRunningTime="2025-10-13 09:01:31.014405886 +0000 UTC m=+1016.162281637" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.040825 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzvpp\" (UniqueName: \"kubernetes.io/projected/75095121-6e99-4595-b06f-7440639ef8cc-kube-api-access-xzvpp\") pod \"75095121-6e99-4595-b06f-7440639ef8cc\" (UID: \"75095121-6e99-4595-b06f-7440639ef8cc\") " Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.041241 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqvxx\" (UniqueName: \"kubernetes.io/projected/3fe757e4-3653-4673-af9d-cf9895e28d3e-kube-api-access-cqvxx\") pod \"3fe757e4-3653-4673-af9d-cf9895e28d3e\" (UID: \"3fe757e4-3653-4673-af9d-cf9895e28d3e\") " Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.062151 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75095121-6e99-4595-b06f-7440639ef8cc-kube-api-access-xzvpp" (OuterVolumeSpecName: "kube-api-access-xzvpp") pod "75095121-6e99-4595-b06f-7440639ef8cc" (UID: "75095121-6e99-4595-b06f-7440639ef8cc"). InnerVolumeSpecName "kube-api-access-xzvpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.079122 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fe757e4-3653-4673-af9d-cf9895e28d3e-kube-api-access-cqvxx" (OuterVolumeSpecName: "kube-api-access-cqvxx") pod "3fe757e4-3653-4673-af9d-cf9895e28d3e" (UID: "3fe757e4-3653-4673-af9d-cf9895e28d3e"). InnerVolumeSpecName "kube-api-access-cqvxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.085288 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-hx5gv"] Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.094808 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-hx5gv"] Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.144733 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzvpp\" (UniqueName: \"kubernetes.io/projected/75095121-6e99-4595-b06f-7440639ef8cc-kube-api-access-xzvpp\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.144763 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqvxx\" (UniqueName: \"kubernetes.io/projected/3fe757e4-3653-4673-af9d-cf9895e28d3e-kube-api-access-cqvxx\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:31 crc kubenswrapper[4685]: E1013 09:01:31.208864 4685 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9aa57e1_fe7b_47b9_9f1b_e8c8900aa5a0.slice/crio-afa5be8b3582675c84d7086fd92fad8ff4320fa17cfa09cdae4ea3b950b26806.scope\": RecentStats: unable to find data in memory cache]" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.348346 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 09:01:31 crc kubenswrapper[4685]: E1013 09:01:31.349006 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fe757e4-3653-4673-af9d-cf9895e28d3e" containerName="mariadb-account-create" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.349021 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fe757e4-3653-4673-af9d-cf9895e28d3e" containerName="mariadb-account-create" Oct 13 09:01:31 crc kubenswrapper[4685]: E1013 09:01:31.349046 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75095121-6e99-4595-b06f-7440639ef8cc" containerName="mariadb-account-create" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.349056 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="75095121-6e99-4595-b06f-7440639ef8cc" containerName="mariadb-account-create" Oct 13 09:01:31 crc kubenswrapper[4685]: E1013 09:01:31.349088 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b263770c-536f-4136-a695-a891267a9be6" containerName="init" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.349095 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b263770c-536f-4136-a695-a891267a9be6" containerName="init" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.349255 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b263770c-536f-4136-a695-a891267a9be6" containerName="init" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.349271 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fe757e4-3653-4673-af9d-cf9895e28d3e" containerName="mariadb-account-create" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.349286 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="75095121-6e99-4595-b06f-7440639ef8cc" containerName="mariadb-account-create" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.359719 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.366576 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gt2pf" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.367210 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.367357 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.392209 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.464522 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-scripts\") pod \"glance-default-external-api-0\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.464573 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.464605 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-logs\") pod \"glance-default-external-api-0\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.464625 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.464646 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlg59\" (UniqueName: \"kubernetes.io/projected/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-kube-api-access-qlg59\") pod \"glance-default-external-api-0\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.464839 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.466705 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-config-data\") pod \"glance-default-external-api-0\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.527318 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b263770c-536f-4136-a695-a891267a9be6" path="/var/lib/kubelet/pods/b263770c-536f-4136-a695-a891267a9be6/volumes" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.569511 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-config-data\") pod \"glance-default-external-api-0\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.569595 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-scripts\") pod \"glance-default-external-api-0\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.569624 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.569643 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-logs\") pod \"glance-default-external-api-0\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.569662 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.569680 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlg59\" (UniqueName: \"kubernetes.io/projected/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-kube-api-access-qlg59\") pod \"glance-default-external-api-0\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.569726 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.571401 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.577327 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.583575 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-config-data\") pod \"glance-default-external-api-0\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.592884 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.608701 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-scripts\") pod \"glance-default-external-api-0\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.610793 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-logs\") pod \"glance-default-external-api-0\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.641887 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlg59\" (UniqueName: \"kubernetes.io/projected/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-kube-api-access-qlg59\") pod \"glance-default-external-api-0\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.655374 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.692383 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.718481 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-8hg5b"] Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.785786 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.788114 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.793273 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.821420 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.883605 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.889463 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbbdd5b-a589-45d8-8e94-a910c30e881d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.889538 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fbbdd5b-a589-45d8-8e94-a910c30e881d-logs\") pod \"glance-default-internal-api-0\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.889575 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0fbbdd5b-a589-45d8-8e94-a910c30e881d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.889672 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbbdd5b-a589-45d8-8e94-a910c30e881d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.889694 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbbdd5b-a589-45d8-8e94-a910c30e881d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.889731 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm2rm\" (UniqueName: \"kubernetes.io/projected/0fbbdd5b-a589-45d8-8e94-a910c30e881d-kube-api-access-fm2rm\") pod \"glance-default-internal-api-0\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.889757 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.991138 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-ovsdbserver-sb\") pod \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.991189 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwnf8\" (UniqueName: \"kubernetes.io/projected/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-kube-api-access-hwnf8\") pod \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.991265 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-dns-svc\") pod \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.991335 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-config\") pod \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.991353 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-dns-swift-storage-0\") pod \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.991415 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-ovsdbserver-nb\") pod \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\" (UID: \"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0\") " Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.991729 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbbdd5b-a589-45d8-8e94-a910c30e881d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.991752 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbbdd5b-a589-45d8-8e94-a910c30e881d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.991804 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm2rm\" (UniqueName: \"kubernetes.io/projected/0fbbdd5b-a589-45d8-8e94-a910c30e881d-kube-api-access-fm2rm\") pod \"glance-default-internal-api-0\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.991831 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.991868 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbbdd5b-a589-45d8-8e94-a910c30e881d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.991933 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fbbdd5b-a589-45d8-8e94-a910c30e881d-logs\") pod \"glance-default-internal-api-0\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.991961 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0fbbdd5b-a589-45d8-8e94-a910c30e881d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.992934 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Oct 13 09:01:31 crc kubenswrapper[4685]: I1013 09:01:31.993572 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0fbbdd5b-a589-45d8-8e94-a910c30e881d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.003737 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fbbdd5b-a589-45d8-8e94-a910c30e881d-logs\") pod \"glance-default-internal-api-0\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.061989 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm2rm\" (UniqueName: \"kubernetes.io/projected/0fbbdd5b-a589-45d8-8e94-a910c30e881d-kube-api-access-fm2rm\") pod \"glance-default-internal-api-0\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.071488 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbbdd5b-a589-45d8-8e94-a910c30e881d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.078099 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbbdd5b-a589-45d8-8e94-a910c30e881d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.083018 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbbdd5b-a589-45d8-8e94-a910c30e881d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.083788 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" event={"ID":"778d19c2-2541-40d9-a00c-efb46a746a32","Type":"ContainerStarted","Data":"d5f0bc39baa4772dd60c30384357f016b1c146d66cf132db7e67931f44542c0a"} Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.083945 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-kube-api-access-hwnf8" (OuterVolumeSpecName: "kube-api-access-hwnf8") pod "c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0" (UID: "c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0"). InnerVolumeSpecName "kube-api-access-hwnf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.094522 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwnf8\" (UniqueName: \"kubernetes.io/projected/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-kube-api-access-hwnf8\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.116659 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.135898 4685 generic.go:334] "Generic (PLEG): container finished" podID="c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0" containerID="afa5be8b3582675c84d7086fd92fad8ff4320fa17cfa09cdae4ea3b950b26806" exitCode=0 Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.136880 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.146296 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" event={"ID":"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0","Type":"ContainerDied","Data":"afa5be8b3582675c84d7086fd92fad8ff4320fa17cfa09cdae4ea3b950b26806"} Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.146341 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-ll5t2" event={"ID":"c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0","Type":"ContainerDied","Data":"d9cdce5d38835674053bdee52c49f671b8fb95326570247e88174c832833f6e1"} Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.146361 4685 scope.go:117] "RemoveContainer" containerID="afa5be8b3582675c84d7086fd92fad8ff4320fa17cfa09cdae4ea3b950b26806" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.146582 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2261-account-create-ndcwb" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.151607 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1a3c-account-create-kqvxn" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.223331 4685 scope.go:117] "RemoveContainer" containerID="3aa2c3741b4424b20ec1e164a1ce790ea7358079b0d22bc17e17e025af91175f" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.232093 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0" (UID: "c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.269372 4685 scope.go:117] "RemoveContainer" containerID="afa5be8b3582675c84d7086fd92fad8ff4320fa17cfa09cdae4ea3b950b26806" Oct 13 09:01:32 crc kubenswrapper[4685]: E1013 09:01:32.275055 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afa5be8b3582675c84d7086fd92fad8ff4320fa17cfa09cdae4ea3b950b26806\": container with ID starting with afa5be8b3582675c84d7086fd92fad8ff4320fa17cfa09cdae4ea3b950b26806 not found: ID does not exist" containerID="afa5be8b3582675c84d7086fd92fad8ff4320fa17cfa09cdae4ea3b950b26806" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.275091 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afa5be8b3582675c84d7086fd92fad8ff4320fa17cfa09cdae4ea3b950b26806"} err="failed to get container status \"afa5be8b3582675c84d7086fd92fad8ff4320fa17cfa09cdae4ea3b950b26806\": rpc error: code = NotFound desc = could not find container \"afa5be8b3582675c84d7086fd92fad8ff4320fa17cfa09cdae4ea3b950b26806\": container with ID starting with afa5be8b3582675c84d7086fd92fad8ff4320fa17cfa09cdae4ea3b950b26806 not found: ID does not exist" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.275112 4685 scope.go:117] "RemoveContainer" containerID="3aa2c3741b4424b20ec1e164a1ce790ea7358079b0d22bc17e17e025af91175f" Oct 13 09:01:32 crc kubenswrapper[4685]: E1013 09:01:32.282558 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3aa2c3741b4424b20ec1e164a1ce790ea7358079b0d22bc17e17e025af91175f\": container with ID starting with 3aa2c3741b4424b20ec1e164a1ce790ea7358079b0d22bc17e17e025af91175f not found: ID does not exist" containerID="3aa2c3741b4424b20ec1e164a1ce790ea7358079b0d22bc17e17e025af91175f" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.282592 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3aa2c3741b4424b20ec1e164a1ce790ea7358079b0d22bc17e17e025af91175f"} err="failed to get container status \"3aa2c3741b4424b20ec1e164a1ce790ea7358079b0d22bc17e17e025af91175f\": rpc error: code = NotFound desc = could not find container \"3aa2c3741b4424b20ec1e164a1ce790ea7358079b0d22bc17e17e025af91175f\": container with ID starting with 3aa2c3741b4424b20ec1e164a1ce790ea7358079b0d22bc17e17e025af91175f not found: ID does not exist" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.290820 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0" (UID: "c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.290840 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0" (UID: "c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.297826 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.297883 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.297897 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.316669 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-config" (OuterVolumeSpecName: "config") pod "c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0" (UID: "c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.321881 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0" (UID: "c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.399257 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-config\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.399287 4685 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.421794 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.492722 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-ll5t2"] Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.499965 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-ll5t2"] Oct 13 09:01:32 crc kubenswrapper[4685]: I1013 09:01:32.669474 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 09:01:33 crc kubenswrapper[4685]: I1013 09:01:33.152114 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"364ae61a-ae8c-4906-92c5-03f4e6d2b66d","Type":"ContainerStarted","Data":"53450170d06c88ff637a50a1231cbe226eb3062cd4ea3b751ab59b147e8b0664"} Oct 13 09:01:33 crc kubenswrapper[4685]: I1013 09:01:33.163806 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" event={"ID":"778d19c2-2541-40d9-a00c-efb46a746a32","Type":"ContainerDied","Data":"8fbbf169b94346ba341c373b92ee4c69f7cf5cb3eeaf4f184ceba78ab8924481"} Oct 13 09:01:33 crc kubenswrapper[4685]: I1013 09:01:33.163839 4685 generic.go:334] "Generic (PLEG): container finished" podID="778d19c2-2541-40d9-a00c-efb46a746a32" containerID="8fbbf169b94346ba341c373b92ee4c69f7cf5cb3eeaf4f184ceba78ab8924481" exitCode=0 Oct 13 09:01:33 crc kubenswrapper[4685]: I1013 09:01:33.239187 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 09:01:33 crc kubenswrapper[4685]: I1013 09:01:33.451095 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:01:33 crc kubenswrapper[4685]: I1013 09:01:33.535215 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0" path="/var/lib/kubelet/pods/c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0/volumes" Oct 13 09:01:34 crc kubenswrapper[4685]: I1013 09:01:34.259024 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" event={"ID":"778d19c2-2541-40d9-a00c-efb46a746a32","Type":"ContainerStarted","Data":"1ee88d52907723bc1c8a7413e7086126a5b37a8ce4fa2dceb8104a586c54fac2"} Oct 13 09:01:34 crc kubenswrapper[4685]: I1013 09:01:34.259362 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:01:34 crc kubenswrapper[4685]: I1013 09:01:34.270825 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0fbbdd5b-a589-45d8-8e94-a910c30e881d","Type":"ContainerStarted","Data":"8ae698f05cc282caf9aee6e7501ae12d717bf96ba8da9aeca06eae909063e217"} Oct 13 09:01:34 crc kubenswrapper[4685]: I1013 09:01:34.272427 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"364ae61a-ae8c-4906-92c5-03f4e6d2b66d","Type":"ContainerStarted","Data":"a03f2fa76c17f79dfff2b3478541b0c82347b7ecebaec4d48aed9af0a6571f48"} Oct 13 09:01:34 crc kubenswrapper[4685]: I1013 09:01:34.296206 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" podStartSLOduration=4.296187166 podStartE2EDuration="4.296187166s" podCreationTimestamp="2025-10-13 09:01:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:01:34.293587465 +0000 UTC m=+1019.441463226" watchObservedRunningTime="2025-10-13 09:01:34.296187166 +0000 UTC m=+1019.444062927" Oct 13 09:01:35 crc kubenswrapper[4685]: I1013 09:01:35.289237 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0fbbdd5b-a589-45d8-8e94-a910c30e881d","Type":"ContainerStarted","Data":"7cd4d421de4acfb1988eccbb34b7a0846c73a74d5e3d333d4344711fbe48d050"} Oct 13 09:01:35 crc kubenswrapper[4685]: I1013 09:01:35.293216 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"364ae61a-ae8c-4906-92c5-03f4e6d2b66d","Type":"ContainerStarted","Data":"3c656ca9c4b8ae9a6ea090bab3b690e12dc4043d1ea1295bb49ef2214196361a"} Oct 13 09:01:35 crc kubenswrapper[4685]: I1013 09:01:35.565191 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.565169404 podStartE2EDuration="5.565169404s" podCreationTimestamp="2025-10-13 09:01:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:01:35.314754448 +0000 UTC m=+1020.462630209" watchObservedRunningTime="2025-10-13 09:01:35.565169404 +0000 UTC m=+1020.713045165" Oct 13 09:01:35 crc kubenswrapper[4685]: I1013 09:01:35.910541 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-7h6qb"] Oct 13 09:01:35 crc kubenswrapper[4685]: E1013 09:01:35.910951 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0" containerName="dnsmasq-dns" Oct 13 09:01:35 crc kubenswrapper[4685]: I1013 09:01:35.910975 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0" containerName="dnsmasq-dns" Oct 13 09:01:35 crc kubenswrapper[4685]: E1013 09:01:35.910991 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0" containerName="init" Oct 13 09:01:35 crc kubenswrapper[4685]: I1013 09:01:35.910999 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0" containerName="init" Oct 13 09:01:35 crc kubenswrapper[4685]: I1013 09:01:35.911200 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9aa57e1-fe7b-47b9-9f1b-e8c8900aa5a0" containerName="dnsmasq-dns" Oct 13 09:01:35 crc kubenswrapper[4685]: I1013 09:01:35.911898 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7h6qb" Oct 13 09:01:35 crc kubenswrapper[4685]: I1013 09:01:35.916524 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 13 09:01:35 crc kubenswrapper[4685]: I1013 09:01:35.916725 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 13 09:01:35 crc kubenswrapper[4685]: I1013 09:01:35.916904 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-9966b" Oct 13 09:01:35 crc kubenswrapper[4685]: I1013 09:01:35.922445 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7h6qb"] Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.109472 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-db-sync-config-data\") pod \"cinder-db-sync-7h6qb\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " pod="openstack/cinder-db-sync-7h6qb" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.109533 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-config-data\") pod \"cinder-db-sync-7h6qb\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " pod="openstack/cinder-db-sync-7h6qb" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.109567 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-scripts\") pod \"cinder-db-sync-7h6qb\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " pod="openstack/cinder-db-sync-7h6qb" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.109581 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-combined-ca-bundle\") pod \"cinder-db-sync-7h6qb\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " pod="openstack/cinder-db-sync-7h6qb" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.109656 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmq9p\" (UniqueName: \"kubernetes.io/projected/b225d240-40a0-4d59-aa19-6355515e9f8c-kube-api-access-fmq9p\") pod \"cinder-db-sync-7h6qb\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " pod="openstack/cinder-db-sync-7h6qb" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.109983 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b225d240-40a0-4d59-aa19-6355515e9f8c-etc-machine-id\") pod \"cinder-db-sync-7h6qb\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " pod="openstack/cinder-db-sync-7h6qb" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.194506 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-5p6fr"] Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.196820 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5p6fr" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.201573 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-b4jjn" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.213061 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmq9p\" (UniqueName: \"kubernetes.io/projected/b225d240-40a0-4d59-aa19-6355515e9f8c-kube-api-access-fmq9p\") pod \"cinder-db-sync-7h6qb\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " pod="openstack/cinder-db-sync-7h6qb" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.213139 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b225d240-40a0-4d59-aa19-6355515e9f8c-etc-machine-id\") pod \"cinder-db-sync-7h6qb\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " pod="openstack/cinder-db-sync-7h6qb" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.213232 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-db-sync-config-data\") pod \"cinder-db-sync-7h6qb\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " pod="openstack/cinder-db-sync-7h6qb" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.213303 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-config-data\") pod \"cinder-db-sync-7h6qb\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " pod="openstack/cinder-db-sync-7h6qb" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.213360 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-combined-ca-bundle\") pod \"cinder-db-sync-7h6qb\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " pod="openstack/cinder-db-sync-7h6qb" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.213388 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-scripts\") pod \"cinder-db-sync-7h6qb\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " pod="openstack/cinder-db-sync-7h6qb" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.215871 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b225d240-40a0-4d59-aa19-6355515e9f8c-etc-machine-id\") pod \"cinder-db-sync-7h6qb\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " pod="openstack/cinder-db-sync-7h6qb" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.216990 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-5p6fr"] Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.230585 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.234382 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-config-data\") pod \"cinder-db-sync-7h6qb\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " pod="openstack/cinder-db-sync-7h6qb" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.242655 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-db-sync-config-data\") pod \"cinder-db-sync-7h6qb\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " pod="openstack/cinder-db-sync-7h6qb" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.248512 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-scripts\") pod \"cinder-db-sync-7h6qb\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " pod="openstack/cinder-db-sync-7h6qb" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.253393 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmq9p\" (UniqueName: \"kubernetes.io/projected/b225d240-40a0-4d59-aa19-6355515e9f8c-kube-api-access-fmq9p\") pod \"cinder-db-sync-7h6qb\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " pod="openstack/cinder-db-sync-7h6qb" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.267070 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-combined-ca-bundle\") pod \"cinder-db-sync-7h6qb\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " pod="openstack/cinder-db-sync-7h6qb" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.315613 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l44f9\" (UniqueName: \"kubernetes.io/projected/05506d41-5449-4eb6-86ce-0ac5f5c880ba-kube-api-access-l44f9\") pod \"barbican-db-sync-5p6fr\" (UID: \"05506d41-5449-4eb6-86ce-0ac5f5c880ba\") " pod="openstack/barbican-db-sync-5p6fr" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.316322 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05506d41-5449-4eb6-86ce-0ac5f5c880ba-combined-ca-bundle\") pod \"barbican-db-sync-5p6fr\" (UID: \"05506d41-5449-4eb6-86ce-0ac5f5c880ba\") " pod="openstack/barbican-db-sync-5p6fr" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.316456 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/05506d41-5449-4eb6-86ce-0ac5f5c880ba-db-sync-config-data\") pod \"barbican-db-sync-5p6fr\" (UID: \"05506d41-5449-4eb6-86ce-0ac5f5c880ba\") " pod="openstack/barbican-db-sync-5p6fr" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.350523 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-9msgq"] Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.351661 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9msgq" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.357854 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-6hrl6" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.358032 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.358305 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.374314 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-9msgq"] Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.420130 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l44f9\" (UniqueName: \"kubernetes.io/projected/05506d41-5449-4eb6-86ce-0ac5f5c880ba-kube-api-access-l44f9\") pod \"barbican-db-sync-5p6fr\" (UID: \"05506d41-5449-4eb6-86ce-0ac5f5c880ba\") " pod="openstack/barbican-db-sync-5p6fr" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.420199 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05506d41-5449-4eb6-86ce-0ac5f5c880ba-combined-ca-bundle\") pod \"barbican-db-sync-5p6fr\" (UID: \"05506d41-5449-4eb6-86ce-0ac5f5c880ba\") " pod="openstack/barbican-db-sync-5p6fr" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.420234 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/05506d41-5449-4eb6-86ce-0ac5f5c880ba-db-sync-config-data\") pod \"barbican-db-sync-5p6fr\" (UID: \"05506d41-5449-4eb6-86ce-0ac5f5c880ba\") " pod="openstack/barbican-db-sync-5p6fr" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.428891 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/05506d41-5449-4eb6-86ce-0ac5f5c880ba-db-sync-config-data\") pod \"barbican-db-sync-5p6fr\" (UID: \"05506d41-5449-4eb6-86ce-0ac5f5c880ba\") " pod="openstack/barbican-db-sync-5p6fr" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.437875 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l44f9\" (UniqueName: \"kubernetes.io/projected/05506d41-5449-4eb6-86ce-0ac5f5c880ba-kube-api-access-l44f9\") pod \"barbican-db-sync-5p6fr\" (UID: \"05506d41-5449-4eb6-86ce-0ac5f5c880ba\") " pod="openstack/barbican-db-sync-5p6fr" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.443002 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05506d41-5449-4eb6-86ce-0ac5f5c880ba-combined-ca-bundle\") pod \"barbican-db-sync-5p6fr\" (UID: \"05506d41-5449-4eb6-86ce-0ac5f5c880ba\") " pod="openstack/barbican-db-sync-5p6fr" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.521583 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee75d00f-3165-45a5-854d-1045b43f2c87-combined-ca-bundle\") pod \"neutron-db-sync-9msgq\" (UID: \"ee75d00f-3165-45a5-854d-1045b43f2c87\") " pod="openstack/neutron-db-sync-9msgq" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.521981 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rppfh\" (UniqueName: \"kubernetes.io/projected/ee75d00f-3165-45a5-854d-1045b43f2c87-kube-api-access-rppfh\") pod \"neutron-db-sync-9msgq\" (UID: \"ee75d00f-3165-45a5-854d-1045b43f2c87\") " pod="openstack/neutron-db-sync-9msgq" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.522190 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ee75d00f-3165-45a5-854d-1045b43f2c87-config\") pod \"neutron-db-sync-9msgq\" (UID: \"ee75d00f-3165-45a5-854d-1045b43f2c87\") " pod="openstack/neutron-db-sync-9msgq" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.544450 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7h6qb" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.615232 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5p6fr" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.623787 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ee75d00f-3165-45a5-854d-1045b43f2c87-config\") pod \"neutron-db-sync-9msgq\" (UID: \"ee75d00f-3165-45a5-854d-1045b43f2c87\") " pod="openstack/neutron-db-sync-9msgq" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.623933 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee75d00f-3165-45a5-854d-1045b43f2c87-combined-ca-bundle\") pod \"neutron-db-sync-9msgq\" (UID: \"ee75d00f-3165-45a5-854d-1045b43f2c87\") " pod="openstack/neutron-db-sync-9msgq" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.623981 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rppfh\" (UniqueName: \"kubernetes.io/projected/ee75d00f-3165-45a5-854d-1045b43f2c87-kube-api-access-rppfh\") pod \"neutron-db-sync-9msgq\" (UID: \"ee75d00f-3165-45a5-854d-1045b43f2c87\") " pod="openstack/neutron-db-sync-9msgq" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.627421 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee75d00f-3165-45a5-854d-1045b43f2c87-combined-ca-bundle\") pod \"neutron-db-sync-9msgq\" (UID: \"ee75d00f-3165-45a5-854d-1045b43f2c87\") " pod="openstack/neutron-db-sync-9msgq" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.627576 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ee75d00f-3165-45a5-854d-1045b43f2c87-config\") pod \"neutron-db-sync-9msgq\" (UID: \"ee75d00f-3165-45a5-854d-1045b43f2c87\") " pod="openstack/neutron-db-sync-9msgq" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.644281 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rppfh\" (UniqueName: \"kubernetes.io/projected/ee75d00f-3165-45a5-854d-1045b43f2c87-kube-api-access-rppfh\") pod \"neutron-db-sync-9msgq\" (UID: \"ee75d00f-3165-45a5-854d-1045b43f2c87\") " pod="openstack/neutron-db-sync-9msgq" Oct 13 09:01:36 crc kubenswrapper[4685]: I1013 09:01:36.675656 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9msgq" Oct 13 09:01:38 crc kubenswrapper[4685]: I1013 09:01:38.336104 4685 generic.go:334] "Generic (PLEG): container finished" podID="8b23db04-161e-4736-85de-a2173dfa0d4f" containerID="09eb66b97da1f79efa9078a0e2afd34299176faadf74af83bf9b0b03d570003f" exitCode=0 Oct 13 09:01:38 crc kubenswrapper[4685]: I1013 09:01:38.336393 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-64vzp" event={"ID":"8b23db04-161e-4736-85de-a2173dfa0d4f","Type":"ContainerDied","Data":"09eb66b97da1f79efa9078a0e2afd34299176faadf74af83bf9b0b03d570003f"} Oct 13 09:01:38 crc kubenswrapper[4685]: I1013 09:01:38.574296 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 09:01:38 crc kubenswrapper[4685]: I1013 09:01:38.574535 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="364ae61a-ae8c-4906-92c5-03f4e6d2b66d" containerName="glance-log" containerID="cri-o://a03f2fa76c17f79dfff2b3478541b0c82347b7ecebaec4d48aed9af0a6571f48" gracePeriod=30 Oct 13 09:01:38 crc kubenswrapper[4685]: I1013 09:01:38.579118 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="364ae61a-ae8c-4906-92c5-03f4e6d2b66d" containerName="glance-httpd" containerID="cri-o://3c656ca9c4b8ae9a6ea090bab3b690e12dc4043d1ea1295bb49ef2214196361a" gracePeriod=30 Oct 13 09:01:38 crc kubenswrapper[4685]: I1013 09:01:38.663779 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 09:01:38 crc kubenswrapper[4685]: I1013 09:01:38.886278 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-695cd469df-pgg2b"] Oct 13 09:01:38 crc kubenswrapper[4685]: I1013 09:01:38.930578 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-db976b79d-q446d"] Oct 13 09:01:38 crc kubenswrapper[4685]: I1013 09:01:38.935719 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:38 crc kubenswrapper[4685]: I1013 09:01:38.942244 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 13 09:01:38 crc kubenswrapper[4685]: I1013 09:01:38.959811 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-db976b79d-q446d"] Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.026641 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7bdb58b6d9-xrlmx"] Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.074092 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-64d69c974d-22pqc"] Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.075454 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.086004 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-combined-ca-bundle\") pod \"horizon-db976b79d-q446d\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.086234 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-config-data\") pod \"horizon-db976b79d-q446d\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.086486 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54czh\" (UniqueName: \"kubernetes.io/projected/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-kube-api-access-54czh\") pod \"horizon-db976b79d-q446d\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.086877 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-scripts\") pod \"horizon-db976b79d-q446d\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.086927 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-horizon-tls-certs\") pod \"horizon-db976b79d-q446d\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.086982 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-horizon-secret-key\") pod \"horizon-db976b79d-q446d\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.087034 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-logs\") pod \"horizon-db976b79d-q446d\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.101469 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64d69c974d-22pqc"] Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.188640 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/126e6e95-80d8-4e68-a541-8e53a4f5b862-scripts\") pod \"horizon-64d69c974d-22pqc\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.188724 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/126e6e95-80d8-4e68-a541-8e53a4f5b862-logs\") pod \"horizon-64d69c974d-22pqc\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.188769 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-scripts\") pod \"horizon-db976b79d-q446d\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.188801 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-horizon-tls-certs\") pod \"horizon-db976b79d-q446d\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.188842 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-horizon-secret-key\") pod \"horizon-db976b79d-q446d\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.188865 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/126e6e95-80d8-4e68-a541-8e53a4f5b862-combined-ca-bundle\") pod \"horizon-64d69c974d-22pqc\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.188900 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/126e6e95-80d8-4e68-a541-8e53a4f5b862-horizon-secret-key\") pod \"horizon-64d69c974d-22pqc\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.188941 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/126e6e95-80d8-4e68-a541-8e53a4f5b862-horizon-tls-certs\") pod \"horizon-64d69c974d-22pqc\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.188966 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-logs\") pod \"horizon-db976b79d-q446d\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.188992 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvh94\" (UniqueName: \"kubernetes.io/projected/126e6e95-80d8-4e68-a541-8e53a4f5b862-kube-api-access-mvh94\") pod \"horizon-64d69c974d-22pqc\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.189029 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-combined-ca-bundle\") pod \"horizon-db976b79d-q446d\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.189061 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-config-data\") pod \"horizon-db976b79d-q446d\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.189124 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54czh\" (UniqueName: \"kubernetes.io/projected/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-kube-api-access-54czh\") pod \"horizon-db976b79d-q446d\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.189158 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/126e6e95-80d8-4e68-a541-8e53a4f5b862-config-data\") pod \"horizon-64d69c974d-22pqc\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.192245 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-logs\") pod \"horizon-db976b79d-q446d\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.192348 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-scripts\") pod \"horizon-db976b79d-q446d\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.193442 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-config-data\") pod \"horizon-db976b79d-q446d\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.198943 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-horizon-tls-certs\") pod \"horizon-db976b79d-q446d\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.200451 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-horizon-secret-key\") pod \"horizon-db976b79d-q446d\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.204083 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-combined-ca-bundle\") pod \"horizon-db976b79d-q446d\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.211684 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54czh\" (UniqueName: \"kubernetes.io/projected/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-kube-api-access-54czh\") pod \"horizon-db976b79d-q446d\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.263780 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.291063 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/126e6e95-80d8-4e68-a541-8e53a4f5b862-config-data\") pod \"horizon-64d69c974d-22pqc\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.291116 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/126e6e95-80d8-4e68-a541-8e53a4f5b862-scripts\") pod \"horizon-64d69c974d-22pqc\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.291153 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/126e6e95-80d8-4e68-a541-8e53a4f5b862-logs\") pod \"horizon-64d69c974d-22pqc\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.291192 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/126e6e95-80d8-4e68-a541-8e53a4f5b862-combined-ca-bundle\") pod \"horizon-64d69c974d-22pqc\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.291216 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/126e6e95-80d8-4e68-a541-8e53a4f5b862-horizon-secret-key\") pod \"horizon-64d69c974d-22pqc\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.291234 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/126e6e95-80d8-4e68-a541-8e53a4f5b862-horizon-tls-certs\") pod \"horizon-64d69c974d-22pqc\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.291254 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvh94\" (UniqueName: \"kubernetes.io/projected/126e6e95-80d8-4e68-a541-8e53a4f5b862-kube-api-access-mvh94\") pod \"horizon-64d69c974d-22pqc\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.292306 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/126e6e95-80d8-4e68-a541-8e53a4f5b862-scripts\") pod \"horizon-64d69c974d-22pqc\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.292630 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/126e6e95-80d8-4e68-a541-8e53a4f5b862-config-data\") pod \"horizon-64d69c974d-22pqc\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.293259 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/126e6e95-80d8-4e68-a541-8e53a4f5b862-logs\") pod \"horizon-64d69c974d-22pqc\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.298212 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/126e6e95-80d8-4e68-a541-8e53a4f5b862-combined-ca-bundle\") pod \"horizon-64d69c974d-22pqc\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.308779 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/126e6e95-80d8-4e68-a541-8e53a4f5b862-horizon-secret-key\") pod \"horizon-64d69c974d-22pqc\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.316783 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvh94\" (UniqueName: \"kubernetes.io/projected/126e6e95-80d8-4e68-a541-8e53a4f5b862-kube-api-access-mvh94\") pod \"horizon-64d69c974d-22pqc\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.317795 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/126e6e95-80d8-4e68-a541-8e53a4f5b862-horizon-tls-certs\") pod \"horizon-64d69c974d-22pqc\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.352467 4685 generic.go:334] "Generic (PLEG): container finished" podID="364ae61a-ae8c-4906-92c5-03f4e6d2b66d" containerID="3c656ca9c4b8ae9a6ea090bab3b690e12dc4043d1ea1295bb49ef2214196361a" exitCode=0 Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.352803 4685 generic.go:334] "Generic (PLEG): container finished" podID="364ae61a-ae8c-4906-92c5-03f4e6d2b66d" containerID="a03f2fa76c17f79dfff2b3478541b0c82347b7ecebaec4d48aed9af0a6571f48" exitCode=143 Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.353061 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"364ae61a-ae8c-4906-92c5-03f4e6d2b66d","Type":"ContainerDied","Data":"3c656ca9c4b8ae9a6ea090bab3b690e12dc4043d1ea1295bb49ef2214196361a"} Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.353091 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"364ae61a-ae8c-4906-92c5-03f4e6d2b66d","Type":"ContainerDied","Data":"a03f2fa76c17f79dfff2b3478541b0c82347b7ecebaec4d48aed9af0a6571f48"} Oct 13 09:01:39 crc kubenswrapper[4685]: I1013 09:01:39.406036 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:40 crc kubenswrapper[4685]: I1013 09:01:40.381508 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0fbbdd5b-a589-45d8-8e94-a910c30e881d","Type":"ContainerStarted","Data":"312d9b4422aa3501411046a13d1ae5110bc5911494abda3e4f41df97e3dd4cc2"} Oct 13 09:01:40 crc kubenswrapper[4685]: I1013 09:01:40.382156 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0fbbdd5b-a589-45d8-8e94-a910c30e881d" containerName="glance-log" containerID="cri-o://7cd4d421de4acfb1988eccbb34b7a0846c73a74d5e3d333d4344711fbe48d050" gracePeriod=30 Oct 13 09:01:40 crc kubenswrapper[4685]: I1013 09:01:40.382407 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0fbbdd5b-a589-45d8-8e94-a910c30e881d" containerName="glance-httpd" containerID="cri-o://312d9b4422aa3501411046a13d1ae5110bc5911494abda3e4f41df97e3dd4cc2" gracePeriod=30 Oct 13 09:01:40 crc kubenswrapper[4685]: I1013 09:01:40.415253 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=10.4152333 podStartE2EDuration="10.4152333s" podCreationTimestamp="2025-10-13 09:01:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:01:40.406447081 +0000 UTC m=+1025.554322842" watchObservedRunningTime="2025-10-13 09:01:40.4152333 +0000 UTC m=+1025.563109061" Oct 13 09:01:41 crc kubenswrapper[4685]: I1013 09:01:41.014355 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:01:41 crc kubenswrapper[4685]: I1013 09:01:41.095314 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-7fttg"] Oct 13 09:01:41 crc kubenswrapper[4685]: I1013 09:01:41.097089 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" podUID="13c3d843-ad90-41fc-9f5f-6935eb976489" containerName="dnsmasq-dns" containerID="cri-o://b817b4e33a9c45dbe7eac5187fcb05112d559a3966f81e8d5ce20dbe4d35211f" gracePeriod=10 Oct 13 09:01:41 crc kubenswrapper[4685]: I1013 09:01:41.399553 4685 generic.go:334] "Generic (PLEG): container finished" podID="0fbbdd5b-a589-45d8-8e94-a910c30e881d" containerID="312d9b4422aa3501411046a13d1ae5110bc5911494abda3e4f41df97e3dd4cc2" exitCode=0 Oct 13 09:01:41 crc kubenswrapper[4685]: I1013 09:01:41.399848 4685 generic.go:334] "Generic (PLEG): container finished" podID="0fbbdd5b-a589-45d8-8e94-a910c30e881d" containerID="7cd4d421de4acfb1988eccbb34b7a0846c73a74d5e3d333d4344711fbe48d050" exitCode=143 Oct 13 09:01:41 crc kubenswrapper[4685]: I1013 09:01:41.399894 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0fbbdd5b-a589-45d8-8e94-a910c30e881d","Type":"ContainerDied","Data":"312d9b4422aa3501411046a13d1ae5110bc5911494abda3e4f41df97e3dd4cc2"} Oct 13 09:01:41 crc kubenswrapper[4685]: I1013 09:01:41.399954 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0fbbdd5b-a589-45d8-8e94-a910c30e881d","Type":"ContainerDied","Data":"7cd4d421de4acfb1988eccbb34b7a0846c73a74d5e3d333d4344711fbe48d050"} Oct 13 09:01:41 crc kubenswrapper[4685]: I1013 09:01:41.401615 4685 generic.go:334] "Generic (PLEG): container finished" podID="13c3d843-ad90-41fc-9f5f-6935eb976489" containerID="b817b4e33a9c45dbe7eac5187fcb05112d559a3966f81e8d5ce20dbe4d35211f" exitCode=0 Oct 13 09:01:41 crc kubenswrapper[4685]: I1013 09:01:41.401646 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" event={"ID":"13c3d843-ad90-41fc-9f5f-6935eb976489","Type":"ContainerDied","Data":"b817b4e33a9c45dbe7eac5187fcb05112d559a3966f81e8d5ce20dbe4d35211f"} Oct 13 09:01:45 crc kubenswrapper[4685]: I1013 09:01:45.227727 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" podUID="13c3d843-ad90-41fc-9f5f-6935eb976489" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: connect: connection refused" Oct 13 09:01:47 crc kubenswrapper[4685]: E1013 09:01:47.952622 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 13 09:01:47 crc kubenswrapper[4685]: E1013 09:01:47.953058 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nb9hc5hc7hfdh579h58ch8bh58bh66ch5f6h9h574hd4h5bh5fbh5c8h589h697hc6h597h659h6bhd6h59bh9ch95h586h567h657h55dh58dh5ffq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-stwmz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7bdb58b6d9-xrlmx_openstack(552e9180-01ac-4cbc-9060-2fa6e6e18e9f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 09:01:47 crc kubenswrapper[4685]: E1013 09:01:47.961325 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7bdb58b6d9-xrlmx" podUID="552e9180-01ac-4cbc-9060-2fa6e6e18e9f" Oct 13 09:01:47 crc kubenswrapper[4685]: E1013 09:01:47.982862 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 13 09:01:47 crc kubenswrapper[4685]: E1013 09:01:47.983329 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh8bhfdh67dh5c7h5c9h8ch5f6h644hf6h4h5d4h7dh56bhffh647h55dh657h55dhb5h5bdh647h54dh568hd8h568h5c9h99h5b9h555hffh55dq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-snxdd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-695cd469df-pgg2b_openstack(48006700-a390-415e-97c6-c083512b1fb7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 09:01:47 crc kubenswrapper[4685]: E1013 09:01:47.989879 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-695cd469df-pgg2b" podUID="48006700-a390-415e-97c6-c083512b1fb7" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.204522 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-64vzp" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.215610 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.286784 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.286850 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-fernet-keys\") pod \"8b23db04-161e-4736-85de-a2173dfa0d4f\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.286907 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mztxl\" (UniqueName: \"kubernetes.io/projected/8b23db04-161e-4736-85de-a2173dfa0d4f-kube-api-access-mztxl\") pod \"8b23db04-161e-4736-85de-a2173dfa0d4f\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.286969 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-scripts\") pod \"8b23db04-161e-4736-85de-a2173dfa0d4f\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.287042 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-credential-keys\") pod \"8b23db04-161e-4736-85de-a2173dfa0d4f\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.287074 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-httpd-run\") pod \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.287149 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-combined-ca-bundle\") pod \"8b23db04-161e-4736-85de-a2173dfa0d4f\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.287181 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-combined-ca-bundle\") pod \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.287240 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-config-data\") pod \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.287273 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-logs\") pod \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.287329 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-config-data\") pod \"8b23db04-161e-4736-85de-a2173dfa0d4f\" (UID: \"8b23db04-161e-4736-85de-a2173dfa0d4f\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.287370 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-scripts\") pod \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.287426 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlg59\" (UniqueName: \"kubernetes.io/projected/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-kube-api-access-qlg59\") pod \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\" (UID: \"364ae61a-ae8c-4906-92c5-03f4e6d2b66d\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.293853 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-logs" (OuterVolumeSpecName: "logs") pod "364ae61a-ae8c-4906-92c5-03f4e6d2b66d" (UID: "364ae61a-ae8c-4906-92c5-03f4e6d2b66d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.298498 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "364ae61a-ae8c-4906-92c5-03f4e6d2b66d" (UID: "364ae61a-ae8c-4906-92c5-03f4e6d2b66d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.300241 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b23db04-161e-4736-85de-a2173dfa0d4f-kube-api-access-mztxl" (OuterVolumeSpecName: "kube-api-access-mztxl") pod "8b23db04-161e-4736-85de-a2173dfa0d4f" (UID: "8b23db04-161e-4736-85de-a2173dfa0d4f"). InnerVolumeSpecName "kube-api-access-mztxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.303027 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8b23db04-161e-4736-85de-a2173dfa0d4f" (UID: "8b23db04-161e-4736-85de-a2173dfa0d4f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.303028 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "364ae61a-ae8c-4906-92c5-03f4e6d2b66d" (UID: "364ae61a-ae8c-4906-92c5-03f4e6d2b66d"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.303092 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-scripts" (OuterVolumeSpecName: "scripts") pod "8b23db04-161e-4736-85de-a2173dfa0d4f" (UID: "8b23db04-161e-4736-85de-a2173dfa0d4f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.306622 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-scripts" (OuterVolumeSpecName: "scripts") pod "364ae61a-ae8c-4906-92c5-03f4e6d2b66d" (UID: "364ae61a-ae8c-4906-92c5-03f4e6d2b66d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.307697 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-kube-api-access-qlg59" (OuterVolumeSpecName: "kube-api-access-qlg59") pod "364ae61a-ae8c-4906-92c5-03f4e6d2b66d" (UID: "364ae61a-ae8c-4906-92c5-03f4e6d2b66d"). InnerVolumeSpecName "kube-api-access-qlg59". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.329204 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "8b23db04-161e-4736-85de-a2173dfa0d4f" (UID: "8b23db04-161e-4736-85de-a2173dfa0d4f"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.346248 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8b23db04-161e-4736-85de-a2173dfa0d4f" (UID: "8b23db04-161e-4736-85de-a2173dfa0d4f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.360247 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "364ae61a-ae8c-4906-92c5-03f4e6d2b66d" (UID: "364ae61a-ae8c-4906-92c5-03f4e6d2b66d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.365071 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-config-data" (OuterVolumeSpecName: "config-data") pod "8b23db04-161e-4736-85de-a2173dfa0d4f" (UID: "8b23db04-161e-4736-85de-a2173dfa0d4f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.388831 4685 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.388858 4685 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.389103 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.389117 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.389126 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-logs\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.389134 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.389141 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.389151 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlg59\" (UniqueName: \"kubernetes.io/projected/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-kube-api-access-qlg59\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.389191 4685 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.389202 4685 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.389211 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mztxl\" (UniqueName: \"kubernetes.io/projected/8b23db04-161e-4736-85de-a2173dfa0d4f-kube-api-access-mztxl\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.389219 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b23db04-161e-4736-85de-a2173dfa0d4f-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.410717 4685 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.444558 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-config-data" (OuterVolumeSpecName: "config-data") pod "364ae61a-ae8c-4906-92c5-03f4e6d2b66d" (UID: "364ae61a-ae8c-4906-92c5-03f4e6d2b66d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.469779 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"364ae61a-ae8c-4906-92c5-03f4e6d2b66d","Type":"ContainerDied","Data":"53450170d06c88ff637a50a1231cbe226eb3062cd4ea3b751ab59b147e8b0664"} Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.469841 4685 scope.go:117] "RemoveContainer" containerID="3c656ca9c4b8ae9a6ea090bab3b690e12dc4043d1ea1295bb49ef2214196361a" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.470021 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.480956 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-64vzp" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.484006 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-64vzp" event={"ID":"8b23db04-161e-4736-85de-a2173dfa0d4f","Type":"ContainerDied","Data":"3155b9a33d1ca7af94ce2d6c23557d9403654ec051d08d9e85e3b8b5beb72de6"} Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.484064 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3155b9a33d1ca7af94ce2d6c23557d9403654ec051d08d9e85e3b8b5beb72de6" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.490515 4685 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.490538 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/364ae61a-ae8c-4906-92c5-03f4e6d2b66d-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.508189 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.554658 4685 scope.go:117] "RemoveContainer" containerID="a03f2fa76c17f79dfff2b3478541b0c82347b7ecebaec4d48aed9af0a6571f48" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.591678 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-dns-svc\") pod \"13c3d843-ad90-41fc-9f5f-6935eb976489\" (UID: \"13c3d843-ad90-41fc-9f5f-6935eb976489\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.591795 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c65ws\" (UniqueName: \"kubernetes.io/projected/13c3d843-ad90-41fc-9f5f-6935eb976489-kube-api-access-c65ws\") pod \"13c3d843-ad90-41fc-9f5f-6935eb976489\" (UID: \"13c3d843-ad90-41fc-9f5f-6935eb976489\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.591837 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-config\") pod \"13c3d843-ad90-41fc-9f5f-6935eb976489\" (UID: \"13c3d843-ad90-41fc-9f5f-6935eb976489\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.591860 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-ovsdbserver-sb\") pod \"13c3d843-ad90-41fc-9f5f-6935eb976489\" (UID: \"13c3d843-ad90-41fc-9f5f-6935eb976489\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.591877 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-ovsdbserver-nb\") pod \"13c3d843-ad90-41fc-9f5f-6935eb976489\" (UID: \"13c3d843-ad90-41fc-9f5f-6935eb976489\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.608841 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13c3d843-ad90-41fc-9f5f-6935eb976489-kube-api-access-c65ws" (OuterVolumeSpecName: "kube-api-access-c65ws") pod "13c3d843-ad90-41fc-9f5f-6935eb976489" (UID: "13c3d843-ad90-41fc-9f5f-6935eb976489"). InnerVolumeSpecName "kube-api-access-c65ws". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.625008 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.669954 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.694437 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0fbbdd5b-a589-45d8-8e94-a910c30e881d-httpd-run\") pod \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.694491 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fm2rm\" (UniqueName: \"kubernetes.io/projected/0fbbdd5b-a589-45d8-8e94-a910c30e881d-kube-api-access-fm2rm\") pod \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.694623 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.694645 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbbdd5b-a589-45d8-8e94-a910c30e881d-combined-ca-bundle\") pod \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.694688 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fbbdd5b-a589-45d8-8e94-a910c30e881d-logs\") pod \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.694739 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbbdd5b-a589-45d8-8e94-a910c30e881d-scripts\") pod \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.694803 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbbdd5b-a589-45d8-8e94-a910c30e881d-config-data\") pod \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\" (UID: \"0fbbdd5b-a589-45d8-8e94-a910c30e881d\") " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.695004 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fbbdd5b-a589-45d8-8e94-a910c30e881d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0fbbdd5b-a589-45d8-8e94-a910c30e881d" (UID: "0fbbdd5b-a589-45d8-8e94-a910c30e881d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.695127 4685 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0fbbdd5b-a589-45d8-8e94-a910c30e881d-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.695144 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c65ws\" (UniqueName: \"kubernetes.io/projected/13c3d843-ad90-41fc-9f5f-6935eb976489-kube-api-access-c65ws\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.716462 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fbbdd5b-a589-45d8-8e94-a910c30e881d-scripts" (OuterVolumeSpecName: "scripts") pod "0fbbdd5b-a589-45d8-8e94-a910c30e881d" (UID: "0fbbdd5b-a589-45d8-8e94-a910c30e881d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.716559 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fbbdd5b-a589-45d8-8e94-a910c30e881d-logs" (OuterVolumeSpecName: "logs") pod "0fbbdd5b-a589-45d8-8e94-a910c30e881d" (UID: "0fbbdd5b-a589-45d8-8e94-a910c30e881d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.728050 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.741781 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "13c3d843-ad90-41fc-9f5f-6935eb976489" (UID: "13c3d843-ad90-41fc-9f5f-6935eb976489"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.744660 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "13c3d843-ad90-41fc-9f5f-6935eb976489" (UID: "13c3d843-ad90-41fc-9f5f-6935eb976489"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.749048 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fbbdd5b-a589-45d8-8e94-a910c30e881d-kube-api-access-fm2rm" (OuterVolumeSpecName: "kube-api-access-fm2rm") pod "0fbbdd5b-a589-45d8-8e94-a910c30e881d" (UID: "0fbbdd5b-a589-45d8-8e94-a910c30e881d"). InnerVolumeSpecName "kube-api-access-fm2rm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.757025 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 09:01:48 crc kubenswrapper[4685]: E1013 09:01:48.757880 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="364ae61a-ae8c-4906-92c5-03f4e6d2b66d" containerName="glance-log" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.757893 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="364ae61a-ae8c-4906-92c5-03f4e6d2b66d" containerName="glance-log" Oct 13 09:01:48 crc kubenswrapper[4685]: E1013 09:01:48.757923 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fbbdd5b-a589-45d8-8e94-a910c30e881d" containerName="glance-log" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.757929 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fbbdd5b-a589-45d8-8e94-a910c30e881d" containerName="glance-log" Oct 13 09:01:48 crc kubenswrapper[4685]: E1013 09:01:48.757953 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="364ae61a-ae8c-4906-92c5-03f4e6d2b66d" containerName="glance-httpd" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.757959 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="364ae61a-ae8c-4906-92c5-03f4e6d2b66d" containerName="glance-httpd" Oct 13 09:01:48 crc kubenswrapper[4685]: E1013 09:01:48.757977 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fbbdd5b-a589-45d8-8e94-a910c30e881d" containerName="glance-httpd" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.757986 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fbbdd5b-a589-45d8-8e94-a910c30e881d" containerName="glance-httpd" Oct 13 09:01:48 crc kubenswrapper[4685]: E1013 09:01:48.758002 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13c3d843-ad90-41fc-9f5f-6935eb976489" containerName="dnsmasq-dns" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.758010 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="13c3d843-ad90-41fc-9f5f-6935eb976489" containerName="dnsmasq-dns" Oct 13 09:01:48 crc kubenswrapper[4685]: E1013 09:01:48.758039 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b23db04-161e-4736-85de-a2173dfa0d4f" containerName="keystone-bootstrap" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.758047 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b23db04-161e-4736-85de-a2173dfa0d4f" containerName="keystone-bootstrap" Oct 13 09:01:48 crc kubenswrapper[4685]: E1013 09:01:48.758062 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13c3d843-ad90-41fc-9f5f-6935eb976489" containerName="init" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.758069 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="13c3d843-ad90-41fc-9f5f-6935eb976489" containerName="init" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.758404 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="364ae61a-ae8c-4906-92c5-03f4e6d2b66d" containerName="glance-httpd" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.758429 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fbbdd5b-a589-45d8-8e94-a910c30e881d" containerName="glance-httpd" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.758454 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b23db04-161e-4736-85de-a2173dfa0d4f" containerName="keystone-bootstrap" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.758469 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="364ae61a-ae8c-4906-92c5-03f4e6d2b66d" containerName="glance-log" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.758479 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="13c3d843-ad90-41fc-9f5f-6935eb976489" containerName="dnsmasq-dns" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.758544 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fbbdd5b-a589-45d8-8e94-a910c30e881d" containerName="glance-log" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.779276 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.785686 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.797969 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fm2rm\" (UniqueName: \"kubernetes.io/projected/0fbbdd5b-a589-45d8-8e94-a910c30e881d-kube-api-access-fm2rm\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.798002 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.798013 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.798022 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fbbdd5b-a589-45d8-8e94-a910c30e881d-logs\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.798030 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbbdd5b-a589-45d8-8e94-a910c30e881d-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.798299 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.799731 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "0fbbdd5b-a589-45d8-8e94-a910c30e881d" (UID: "0fbbdd5b-a589-45d8-8e94-a910c30e881d"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.827273 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "13c3d843-ad90-41fc-9f5f-6935eb976489" (UID: "13c3d843-ad90-41fc-9f5f-6935eb976489"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.846650 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fbbdd5b-a589-45d8-8e94-a910c30e881d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0fbbdd5b-a589-45d8-8e94-a910c30e881d" (UID: "0fbbdd5b-a589-45d8-8e94-a910c30e881d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.856847 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-config" (OuterVolumeSpecName: "config") pod "13c3d843-ad90-41fc-9f5f-6935eb976489" (UID: "13c3d843-ad90-41fc-9f5f-6935eb976489"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.864147 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.868948 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-db976b79d-q446d"] Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.899380 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-scripts\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.899579 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xh78\" (UniqueName: \"kubernetes.io/projected/a9cfba62-3301-4ae3-b32c-d542302792d4-kube-api-access-5xh78\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.899754 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a9cfba62-3301-4ae3-b32c-d542302792d4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.899844 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-config-data\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.899959 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.900205 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.900325 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.900375 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9cfba62-3301-4ae3-b32c-d542302792d4-logs\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.900507 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.900525 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c3d843-ad90-41fc-9f5f-6935eb976489-config\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.900545 4685 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.900555 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbbdd5b-a589-45d8-8e94-a910c30e881d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.911492 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fbbdd5b-a589-45d8-8e94-a910c30e881d-config-data" (OuterVolumeSpecName: "config-data") pod "0fbbdd5b-a589-45d8-8e94-a910c30e881d" (UID: "0fbbdd5b-a589-45d8-8e94-a910c30e881d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:48 crc kubenswrapper[4685]: I1013 09:01:48.938084 4685 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.003754 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-9msgq"] Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.004536 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xh78\" (UniqueName: \"kubernetes.io/projected/a9cfba62-3301-4ae3-b32c-d542302792d4-kube-api-access-5xh78\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.007783 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a9cfba62-3301-4ae3-b32c-d542302792d4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.008006 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-config-data\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.009188 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.009403 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.010039 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.010160 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9cfba62-3301-4ae3-b32c-d542302792d4-logs\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.010273 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-scripts\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.010387 4685 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.019272 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a9cfba62-3301-4ae3-b32c-d542302792d4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.019563 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.025576 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9cfba62-3301-4ae3-b32c-d542302792d4-logs\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.032420 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbbdd5b-a589-45d8-8e94-a910c30e881d-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.046545 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.047133 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.048461 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7h6qb"] Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.052814 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-scripts\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.053436 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xh78\" (UniqueName: \"kubernetes.io/projected/a9cfba62-3301-4ae3-b32c-d542302792d4-kube-api-access-5xh78\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.100165 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-config-data\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.130390 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " pod="openstack/glance-default-external-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.240668 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bdb58b6d9-xrlmx" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.274423 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64d69c974d-22pqc"] Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.280588 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-5p6fr"] Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.312806 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-695cd469df-pgg2b" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.421743 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.438774 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48006700-a390-415e-97c6-c083512b1fb7-logs\") pod \"48006700-a390-415e-97c6-c083512b1fb7\" (UID: \"48006700-a390-415e-97c6-c083512b1fb7\") " Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.438934 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snxdd\" (UniqueName: \"kubernetes.io/projected/48006700-a390-415e-97c6-c083512b1fb7-kube-api-access-snxdd\") pod \"48006700-a390-415e-97c6-c083512b1fb7\" (UID: \"48006700-a390-415e-97c6-c083512b1fb7\") " Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.438985 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48006700-a390-415e-97c6-c083512b1fb7-scripts\") pod \"48006700-a390-415e-97c6-c083512b1fb7\" (UID: \"48006700-a390-415e-97c6-c083512b1fb7\") " Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.439006 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/48006700-a390-415e-97c6-c083512b1fb7-config-data\") pod \"48006700-a390-415e-97c6-c083512b1fb7\" (UID: \"48006700-a390-415e-97c6-c083512b1fb7\") " Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.439056 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stwmz\" (UniqueName: \"kubernetes.io/projected/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-kube-api-access-stwmz\") pod \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\" (UID: \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\") " Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.439104 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-scripts\") pod \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\" (UID: \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\") " Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.439127 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-config-data\") pod \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\" (UID: \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\") " Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.439195 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/48006700-a390-415e-97c6-c083512b1fb7-horizon-secret-key\") pod \"48006700-a390-415e-97c6-c083512b1fb7\" (UID: \"48006700-a390-415e-97c6-c083512b1fb7\") " Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.439245 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-horizon-secret-key\") pod \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\" (UID: \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\") " Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.439266 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-logs\") pod \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\" (UID: \"552e9180-01ac-4cbc-9060-2fa6e6e18e9f\") " Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.439791 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-logs" (OuterVolumeSpecName: "logs") pod "552e9180-01ac-4cbc-9060-2fa6e6e18e9f" (UID: "552e9180-01ac-4cbc-9060-2fa6e6e18e9f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.440222 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-scripts" (OuterVolumeSpecName: "scripts") pod "552e9180-01ac-4cbc-9060-2fa6e6e18e9f" (UID: "552e9180-01ac-4cbc-9060-2fa6e6e18e9f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.440229 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48006700-a390-415e-97c6-c083512b1fb7-scripts" (OuterVolumeSpecName: "scripts") pod "48006700-a390-415e-97c6-c083512b1fb7" (UID: "48006700-a390-415e-97c6-c083512b1fb7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.440503 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48006700-a390-415e-97c6-c083512b1fb7-logs" (OuterVolumeSpecName: "logs") pod "48006700-a390-415e-97c6-c083512b1fb7" (UID: "48006700-a390-415e-97c6-c083512b1fb7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.441026 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-config-data" (OuterVolumeSpecName: "config-data") pod "552e9180-01ac-4cbc-9060-2fa6e6e18e9f" (UID: "552e9180-01ac-4cbc-9060-2fa6e6e18e9f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.441031 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48006700-a390-415e-97c6-c083512b1fb7-config-data" (OuterVolumeSpecName: "config-data") pod "48006700-a390-415e-97c6-c083512b1fb7" (UID: "48006700-a390-415e-97c6-c083512b1fb7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.448360 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48006700-a390-415e-97c6-c083512b1fb7-kube-api-access-snxdd" (OuterVolumeSpecName: "kube-api-access-snxdd") pod "48006700-a390-415e-97c6-c083512b1fb7" (UID: "48006700-a390-415e-97c6-c083512b1fb7"). InnerVolumeSpecName "kube-api-access-snxdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.448546 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48006700-a390-415e-97c6-c083512b1fb7-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "48006700-a390-415e-97c6-c083512b1fb7" (UID: "48006700-a390-415e-97c6-c083512b1fb7"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.448773 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-kube-api-access-stwmz" (OuterVolumeSpecName: "kube-api-access-stwmz") pod "552e9180-01ac-4cbc-9060-2fa6e6e18e9f" (UID: "552e9180-01ac-4cbc-9060-2fa6e6e18e9f"). InnerVolumeSpecName "kube-api-access-stwmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.449081 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "552e9180-01ac-4cbc-9060-2fa6e6e18e9f" (UID: "552e9180-01ac-4cbc-9060-2fa6e6e18e9f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.499903 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4phvd" event={"ID":"c5afcbfe-3193-426b-97fe-4e43f4a483d8","Type":"ContainerStarted","Data":"d9b0e879150b37dc976a9e934abaa9d8257375d68d2c2ac8dc3ac813e49f3cd8"} Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.504129 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bdb58b6d9-xrlmx" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.527012 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-4phvd" podStartSLOduration=4.189534681 podStartE2EDuration="23.526993695s" podCreationTimestamp="2025-10-13 09:01:26 +0000 UTC" firstStartedPulling="2025-10-13 09:01:28.748676631 +0000 UTC m=+1013.896552392" lastFinishedPulling="2025-10-13 09:01:48.086135645 +0000 UTC m=+1033.234011406" observedRunningTime="2025-10-13 09:01:49.522992195 +0000 UTC m=+1034.670867956" watchObservedRunningTime="2025-10-13 09:01:49.526993695 +0000 UTC m=+1034.674869456" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.531199 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.540401 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.540426 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.540436 4685 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/48006700-a390-415e-97c6-c083512b1fb7-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.540444 4685 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.540452 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-logs\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.540459 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48006700-a390-415e-97c6-c083512b1fb7-logs\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.540466 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snxdd\" (UniqueName: \"kubernetes.io/projected/48006700-a390-415e-97c6-c083512b1fb7-kube-api-access-snxdd\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.540475 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48006700-a390-415e-97c6-c083512b1fb7-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.540482 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/48006700-a390-415e-97c6-c083512b1fb7-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.540490 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stwmz\" (UniqueName: \"kubernetes.io/projected/552e9180-01ac-4cbc-9060-2fa6e6e18e9f-kube-api-access-stwmz\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.556373 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="364ae61a-ae8c-4906-92c5-03f4e6d2b66d" path="/var/lib/kubelet/pods/364ae61a-ae8c-4906-92c5-03f4e6d2b66d/volumes" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.567418 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bdb58b6d9-xrlmx" event={"ID":"552e9180-01ac-4cbc-9060-2fa6e6e18e9f","Type":"ContainerDied","Data":"eeb7340908b4e870a0499fad831954f05d8f9f158c9a7687cfb1ae1cc6cc898f"} Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.567450 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-db976b79d-q446d" event={"ID":"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc","Type":"ContainerStarted","Data":"2a683f973e75bb1e5041098aa03edac5179ea0a65e33a3ad0f90f6afc26d2f2d"} Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.567487 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7h6qb" event={"ID":"b225d240-40a0-4d59-aa19-6355515e9f8c","Type":"ContainerStarted","Data":"6c9cd308010bd180f39154e0de4b0369c02187761d7e1ecfa996b4b513a87b6b"} Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.567499 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0fbbdd5b-a589-45d8-8e94-a910c30e881d","Type":"ContainerDied","Data":"8ae698f05cc282caf9aee6e7501ae12d717bf96ba8da9aeca06eae909063e217"} Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.567511 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9msgq" event={"ID":"ee75d00f-3165-45a5-854d-1045b43f2c87","Type":"ContainerStarted","Data":"ee542b97c2ea580a1848390372ad4dd4d87e49fcb174c14691b7a9017f8a1a5c"} Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.567520 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9msgq" event={"ID":"ee75d00f-3165-45a5-854d-1045b43f2c87","Type":"ContainerStarted","Data":"d9f5d6878800827b2e5dcbd3db3436e3ac03c905078486a523d045f2ed50a2eb"} Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.573673 4685 scope.go:117] "RemoveContainer" containerID="312d9b4422aa3501411046a13d1ae5110bc5911494abda3e4f41df97e3dd4cc2" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.577090 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" event={"ID":"13c3d843-ad90-41fc-9f5f-6935eb976489","Type":"ContainerDied","Data":"684d8eb91a1de29de2393f631e657a8894b63dbdc6da4fc5914ff63d1f4e86b4"} Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.577121 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-7fttg" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.579897 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7bdb58b6d9-xrlmx"] Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.589332 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7bdb58b6d9-xrlmx"] Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.591866 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64d69c974d-22pqc" event={"ID":"126e6e95-80d8-4e68-a541-8e53a4f5b862","Type":"ContainerStarted","Data":"34092f0c3725fb957baa150ccc4eaa3c17784f4348ba496fa3ee32ff339fac93"} Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.597399 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-9msgq" podStartSLOduration=13.597386736 podStartE2EDuration="13.597386736s" podCreationTimestamp="2025-10-13 09:01:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:01:49.592540544 +0000 UTC m=+1034.740416295" watchObservedRunningTime="2025-10-13 09:01:49.597386736 +0000 UTC m=+1034.745262497" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.603904 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5p6fr" event={"ID":"05506d41-5449-4eb6-86ce-0ac5f5c880ba","Type":"ContainerStarted","Data":"8e9e8ef00358a7eabc8d68ddaf92adcb440ded6caecc8df0037c974b88e20885"} Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.609686 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-695cd469df-pgg2b" event={"ID":"48006700-a390-415e-97c6-c083512b1fb7","Type":"ContainerDied","Data":"c581ae18eaadd387e55dadb02db33cfb45a858754ce32a46ebacb32c2f93901b"} Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.609725 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-695cd469df-pgg2b" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.615298 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ffc408c-86d3-4e93-9d49-6ef39cb7d959","Type":"ContainerStarted","Data":"a392b4d4475d36fb966ec8ba4fcd6cb1bd6f904281b206bd0573f0b1765e6cf7"} Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.633561 4685 scope.go:117] "RemoveContainer" containerID="7cd4d421de4acfb1988eccbb34b7a0846c73a74d5e3d333d4344711fbe48d050" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.755906 4685 scope.go:117] "RemoveContainer" containerID="b817b4e33a9c45dbe7eac5187fcb05112d559a3966f81e8d5ce20dbe4d35211f" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.757091 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.781079 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.822990 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.824633 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.832154 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.832348 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.848233 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.879063 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-7fttg"] Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.951433 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.951523 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af10022e-b9b5-4bed-8c80-9064b711f837-logs\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.951597 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.951617 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-config-data\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.951658 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/af10022e-b9b5-4bed-8c80-9064b711f837-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.951698 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hpdn\" (UniqueName: \"kubernetes.io/projected/af10022e-b9b5-4bed-8c80-9064b711f837-kube-api-access-5hpdn\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.951739 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-scripts\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.951769 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.998181 4685 scope.go:117] "RemoveContainer" containerID="c834de30ea8b9cd3a8cb84f6f316c5188d2ff326b7a4d20aa5123e9f211bea84" Oct 13 09:01:49 crc kubenswrapper[4685]: I1013 09:01:49.999938 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-7fttg"] Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.035435 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-695cd469df-pgg2b"] Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.048653 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-695cd469df-pgg2b"] Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.053776 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.053832 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af10022e-b9b5-4bed-8c80-9064b711f837-logs\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.053893 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.053931 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-config-data\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.053971 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/af10022e-b9b5-4bed-8c80-9064b711f837-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.053992 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hpdn\" (UniqueName: \"kubernetes.io/projected/af10022e-b9b5-4bed-8c80-9064b711f837-kube-api-access-5hpdn\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.054018 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-scripts\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.054059 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.055769 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/af10022e-b9b5-4bed-8c80-9064b711f837-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.055793 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af10022e-b9b5-4bed-8c80-9064b711f837-logs\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.056173 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.067421 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.067638 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-scripts\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.076882 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.088167 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-config-data\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.091474 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hpdn\" (UniqueName: \"kubernetes.io/projected/af10022e-b9b5-4bed-8c80-9064b711f837-kube-api-access-5hpdn\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.213291 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.312382 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.503125 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.655063 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a9cfba62-3301-4ae3-b32c-d542302792d4","Type":"ContainerStarted","Data":"fa655b2dfe5494ba5b83f8c375a3bb91ead32f8e8f09769d1ec1ff604aaa9ac1"} Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.688148 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64d69c974d-22pqc" event={"ID":"126e6e95-80d8-4e68-a541-8e53a4f5b862","Type":"ContainerStarted","Data":"057d5f72232fe1033e30b206d78832bccc56b3093599e6d33773993e1467b3c1"} Oct 13 09:01:50 crc kubenswrapper[4685]: I1013 09:01:50.711197 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-db976b79d-q446d" event={"ID":"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc","Type":"ContainerStarted","Data":"53e463eff1a6fb595b65a78ef21731ec4ee12f240e505e924644e92f36c5e9c4"} Oct 13 09:01:51 crc kubenswrapper[4685]: I1013 09:01:51.184357 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 09:01:51 crc kubenswrapper[4685]: I1013 09:01:51.520711 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fbbdd5b-a589-45d8-8e94-a910c30e881d" path="/var/lib/kubelet/pods/0fbbdd5b-a589-45d8-8e94-a910c30e881d/volumes" Oct 13 09:01:51 crc kubenswrapper[4685]: I1013 09:01:51.521725 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13c3d843-ad90-41fc-9f5f-6935eb976489" path="/var/lib/kubelet/pods/13c3d843-ad90-41fc-9f5f-6935eb976489/volumes" Oct 13 09:01:51 crc kubenswrapper[4685]: I1013 09:01:51.522421 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48006700-a390-415e-97c6-c083512b1fb7" path="/var/lib/kubelet/pods/48006700-a390-415e-97c6-c083512b1fb7/volumes" Oct 13 09:01:51 crc kubenswrapper[4685]: I1013 09:01:51.530540 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="552e9180-01ac-4cbc-9060-2fa6e6e18e9f" path="/var/lib/kubelet/pods/552e9180-01ac-4cbc-9060-2fa6e6e18e9f/volumes" Oct 13 09:01:51 crc kubenswrapper[4685]: I1013 09:01:51.732652 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a9cfba62-3301-4ae3-b32c-d542302792d4","Type":"ContainerStarted","Data":"b645002af7b6e1599aca65be6d1ea012273493a144aa47de695155e939648e51"} Oct 13 09:01:51 crc kubenswrapper[4685]: I1013 09:01:51.734992 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64d69c974d-22pqc" event={"ID":"126e6e95-80d8-4e68-a541-8e53a4f5b862","Type":"ContainerStarted","Data":"b115917cd44da7ffaccbe91d120ee61d59a4460f53f619a7e3e0215ca2922259"} Oct 13 09:01:51 crc kubenswrapper[4685]: I1013 09:01:51.741598 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-db976b79d-q446d" event={"ID":"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc","Type":"ContainerStarted","Data":"87642f946b3e4dd470d2686255f2a7452ddf07f367a7a524a315e59e389b70f5"} Oct 13 09:01:51 crc kubenswrapper[4685]: I1013 09:01:51.760035 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-64d69c974d-22pqc" podStartSLOduration=12.036400326 podStartE2EDuration="12.760016267s" podCreationTimestamp="2025-10-13 09:01:39 +0000 UTC" firstStartedPulling="2025-10-13 09:01:49.295497826 +0000 UTC m=+1034.443373587" lastFinishedPulling="2025-10-13 09:01:50.019113767 +0000 UTC m=+1035.166989528" observedRunningTime="2025-10-13 09:01:51.756444369 +0000 UTC m=+1036.904320150" watchObservedRunningTime="2025-10-13 09:01:51.760016267 +0000 UTC m=+1036.907892038" Oct 13 09:01:51 crc kubenswrapper[4685]: I1013 09:01:51.793222 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-db976b79d-q446d" podStartSLOduration=13.131882932 podStartE2EDuration="13.793200553s" podCreationTimestamp="2025-10-13 09:01:38 +0000 UTC" firstStartedPulling="2025-10-13 09:01:48.830360289 +0000 UTC m=+1033.978236040" lastFinishedPulling="2025-10-13 09:01:49.4916779 +0000 UTC m=+1034.639553661" observedRunningTime="2025-10-13 09:01:51.783875658 +0000 UTC m=+1036.931751409" watchObservedRunningTime="2025-10-13 09:01:51.793200553 +0000 UTC m=+1036.941076304" Oct 13 09:01:52 crc kubenswrapper[4685]: I1013 09:01:52.758830 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"af10022e-b9b5-4bed-8c80-9064b711f837","Type":"ContainerStarted","Data":"08fe3b449e8aeacd341757b0b6b87e803db87305a67a377cc3bf7c3c04ee6e11"} Oct 13 09:01:52 crc kubenswrapper[4685]: I1013 09:01:52.781519 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ffc408c-86d3-4e93-9d49-6ef39cb7d959","Type":"ContainerStarted","Data":"da0a042d918652a2a33633572db233d9037398e55494df44006a3b65c0286815"} Oct 13 09:01:52 crc kubenswrapper[4685]: I1013 09:01:52.979860 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:01:52 crc kubenswrapper[4685]: I1013 09:01:52.979907 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:01:53 crc kubenswrapper[4685]: I1013 09:01:53.795706 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a9cfba62-3301-4ae3-b32c-d542302792d4","Type":"ContainerStarted","Data":"6faa94ee30712a6d42db9f6e7ed17637643bdf663c088d057871184b5370d738"} Oct 13 09:01:53 crc kubenswrapper[4685]: I1013 09:01:53.796895 4685 generic.go:334] "Generic (PLEG): container finished" podID="c5afcbfe-3193-426b-97fe-4e43f4a483d8" containerID="d9b0e879150b37dc976a9e934abaa9d8257375d68d2c2ac8dc3ac813e49f3cd8" exitCode=0 Oct 13 09:01:53 crc kubenswrapper[4685]: I1013 09:01:53.796951 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4phvd" event={"ID":"c5afcbfe-3193-426b-97fe-4e43f4a483d8","Type":"ContainerDied","Data":"d9b0e879150b37dc976a9e934abaa9d8257375d68d2c2ac8dc3ac813e49f3cd8"} Oct 13 09:01:53 crc kubenswrapper[4685]: I1013 09:01:53.798573 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"af10022e-b9b5-4bed-8c80-9064b711f837","Type":"ContainerStarted","Data":"a5aee9e8730b63472aa83483317267b5ae11e7ad15304edd44922a53a600a0cb"} Oct 13 09:01:53 crc kubenswrapper[4685]: I1013 09:01:53.798601 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"af10022e-b9b5-4bed-8c80-9064b711f837","Type":"ContainerStarted","Data":"99516221fed8e294d62a6cb32d71ed431230ed3bc0d03203306b0083cec2be29"} Oct 13 09:01:53 crc kubenswrapper[4685]: I1013 09:01:53.852204 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.8521863750000005 podStartE2EDuration="5.852186375s" podCreationTimestamp="2025-10-13 09:01:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:01:53.828441196 +0000 UTC m=+1038.976316957" watchObservedRunningTime="2025-10-13 09:01:53.852186375 +0000 UTC m=+1039.000062136" Oct 13 09:01:53 crc kubenswrapper[4685]: I1013 09:01:53.852779 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.852773601 podStartE2EDuration="4.852773601s" podCreationTimestamp="2025-10-13 09:01:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:01:53.850441977 +0000 UTC m=+1038.998317738" watchObservedRunningTime="2025-10-13 09:01:53.852773601 +0000 UTC m=+1039.000649362" Oct 13 09:01:57 crc kubenswrapper[4685]: I1013 09:01:57.406877 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4phvd" Oct 13 09:01:57 crc kubenswrapper[4685]: I1013 09:01:57.520835 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcb7p\" (UniqueName: \"kubernetes.io/projected/c5afcbfe-3193-426b-97fe-4e43f4a483d8-kube-api-access-kcb7p\") pod \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\" (UID: \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\") " Oct 13 09:01:57 crc kubenswrapper[4685]: I1013 09:01:57.520929 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5afcbfe-3193-426b-97fe-4e43f4a483d8-scripts\") pod \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\" (UID: \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\") " Oct 13 09:01:57 crc kubenswrapper[4685]: I1013 09:01:57.520985 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5afcbfe-3193-426b-97fe-4e43f4a483d8-combined-ca-bundle\") pod \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\" (UID: \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\") " Oct 13 09:01:57 crc kubenswrapper[4685]: I1013 09:01:57.521018 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5afcbfe-3193-426b-97fe-4e43f4a483d8-config-data\") pod \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\" (UID: \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\") " Oct 13 09:01:57 crc kubenswrapper[4685]: I1013 09:01:57.521075 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5afcbfe-3193-426b-97fe-4e43f4a483d8-logs\") pod \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\" (UID: \"c5afcbfe-3193-426b-97fe-4e43f4a483d8\") " Oct 13 09:01:57 crc kubenswrapper[4685]: I1013 09:01:57.521749 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5afcbfe-3193-426b-97fe-4e43f4a483d8-logs" (OuterVolumeSpecName: "logs") pod "c5afcbfe-3193-426b-97fe-4e43f4a483d8" (UID: "c5afcbfe-3193-426b-97fe-4e43f4a483d8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:01:57 crc kubenswrapper[4685]: I1013 09:01:57.528085 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5afcbfe-3193-426b-97fe-4e43f4a483d8-scripts" (OuterVolumeSpecName: "scripts") pod "c5afcbfe-3193-426b-97fe-4e43f4a483d8" (UID: "c5afcbfe-3193-426b-97fe-4e43f4a483d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:57 crc kubenswrapper[4685]: I1013 09:01:57.532043 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5afcbfe-3193-426b-97fe-4e43f4a483d8-kube-api-access-kcb7p" (OuterVolumeSpecName: "kube-api-access-kcb7p") pod "c5afcbfe-3193-426b-97fe-4e43f4a483d8" (UID: "c5afcbfe-3193-426b-97fe-4e43f4a483d8"). InnerVolumeSpecName "kube-api-access-kcb7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:01:57 crc kubenswrapper[4685]: I1013 09:01:57.551743 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5afcbfe-3193-426b-97fe-4e43f4a483d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5afcbfe-3193-426b-97fe-4e43f4a483d8" (UID: "c5afcbfe-3193-426b-97fe-4e43f4a483d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:57 crc kubenswrapper[4685]: I1013 09:01:57.595449 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5afcbfe-3193-426b-97fe-4e43f4a483d8-config-data" (OuterVolumeSpecName: "config-data") pod "c5afcbfe-3193-426b-97fe-4e43f4a483d8" (UID: "c5afcbfe-3193-426b-97fe-4e43f4a483d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:01:57 crc kubenswrapper[4685]: I1013 09:01:57.622594 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5afcbfe-3193-426b-97fe-4e43f4a483d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:57 crc kubenswrapper[4685]: I1013 09:01:57.622625 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5afcbfe-3193-426b-97fe-4e43f4a483d8-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:57 crc kubenswrapper[4685]: I1013 09:01:57.622634 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5afcbfe-3193-426b-97fe-4e43f4a483d8-logs\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:57 crc kubenswrapper[4685]: I1013 09:01:57.622643 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcb7p\" (UniqueName: \"kubernetes.io/projected/c5afcbfe-3193-426b-97fe-4e43f4a483d8-kube-api-access-kcb7p\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:57 crc kubenswrapper[4685]: I1013 09:01:57.622652 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5afcbfe-3193-426b-97fe-4e43f4a483d8-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:01:57 crc kubenswrapper[4685]: I1013 09:01:57.840389 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4phvd" event={"ID":"c5afcbfe-3193-426b-97fe-4e43f4a483d8","Type":"ContainerDied","Data":"e2311f145362707a4f3ea678c410be719e938553a3ce91d16679b111b8281f36"} Oct 13 09:01:57 crc kubenswrapper[4685]: I1013 09:01:57.840437 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2311f145362707a4f3ea678c410be719e938553a3ce91d16679b111b8281f36" Oct 13 09:01:57 crc kubenswrapper[4685]: I1013 09:01:57.840500 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4phvd" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.518581 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7d4ff599b8-7thr6"] Oct 13 09:01:58 crc kubenswrapper[4685]: E1013 09:01:58.518936 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5afcbfe-3193-426b-97fe-4e43f4a483d8" containerName="placement-db-sync" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.518947 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5afcbfe-3193-426b-97fe-4e43f4a483d8" containerName="placement-db-sync" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.519100 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5afcbfe-3193-426b-97fe-4e43f4a483d8" containerName="placement-db-sync" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.520139 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.524158 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.524164 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.524354 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.525413 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-cfczb" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.525958 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.556135 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7d4ff599b8-7thr6"] Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.640999 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-scripts\") pod \"placement-7d4ff599b8-7thr6\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.641043 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-combined-ca-bundle\") pod \"placement-7d4ff599b8-7thr6\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.641328 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19f3def1-e242-41ec-8465-164d53d878a4-logs\") pod \"placement-7d4ff599b8-7thr6\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.642100 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq75x\" (UniqueName: \"kubernetes.io/projected/19f3def1-e242-41ec-8465-164d53d878a4-kube-api-access-sq75x\") pod \"placement-7d4ff599b8-7thr6\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.642243 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-internal-tls-certs\") pod \"placement-7d4ff599b8-7thr6\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.642294 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-config-data\") pod \"placement-7d4ff599b8-7thr6\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.642348 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-public-tls-certs\") pod \"placement-7d4ff599b8-7thr6\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.743808 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-scripts\") pod \"placement-7d4ff599b8-7thr6\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.743852 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-combined-ca-bundle\") pod \"placement-7d4ff599b8-7thr6\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.744599 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19f3def1-e242-41ec-8465-164d53d878a4-logs\") pod \"placement-7d4ff599b8-7thr6\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.744661 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq75x\" (UniqueName: \"kubernetes.io/projected/19f3def1-e242-41ec-8465-164d53d878a4-kube-api-access-sq75x\") pod \"placement-7d4ff599b8-7thr6\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.744688 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-internal-tls-certs\") pod \"placement-7d4ff599b8-7thr6\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.744721 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-config-data\") pod \"placement-7d4ff599b8-7thr6\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.744748 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-public-tls-certs\") pod \"placement-7d4ff599b8-7thr6\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.745086 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19f3def1-e242-41ec-8465-164d53d878a4-logs\") pod \"placement-7d4ff599b8-7thr6\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.748532 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-scripts\") pod \"placement-7d4ff599b8-7thr6\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.748845 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-combined-ca-bundle\") pod \"placement-7d4ff599b8-7thr6\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.749902 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-config-data\") pod \"placement-7d4ff599b8-7thr6\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.750840 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-public-tls-certs\") pod \"placement-7d4ff599b8-7thr6\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.752488 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-internal-tls-certs\") pod \"placement-7d4ff599b8-7thr6\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.768471 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq75x\" (UniqueName: \"kubernetes.io/projected/19f3def1-e242-41ec-8465-164d53d878a4-kube-api-access-sq75x\") pod \"placement-7d4ff599b8-7thr6\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:58 crc kubenswrapper[4685]: I1013 09:01:58.857851 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:01:59 crc kubenswrapper[4685]: I1013 09:01:59.264759 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:59 crc kubenswrapper[4685]: I1013 09:01:59.265205 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-db976b79d-q446d" Oct 13 09:01:59 crc kubenswrapper[4685]: I1013 09:01:59.407413 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:59 crc kubenswrapper[4685]: I1013 09:01:59.407737 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:01:59 crc kubenswrapper[4685]: I1013 09:01:59.423486 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 13 09:01:59 crc kubenswrapper[4685]: I1013 09:01:59.423591 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 13 09:01:59 crc kubenswrapper[4685]: I1013 09:01:59.501800 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 13 09:01:59 crc kubenswrapper[4685]: I1013 09:01:59.502465 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 13 09:01:59 crc kubenswrapper[4685]: I1013 09:01:59.861450 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 13 09:01:59 crc kubenswrapper[4685]: I1013 09:01:59.861932 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 13 09:02:00 crc kubenswrapper[4685]: I1013 09:02:00.503570 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 13 09:02:00 crc kubenswrapper[4685]: I1013 09:02:00.503618 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 13 09:02:00 crc kubenswrapper[4685]: I1013 09:02:00.558585 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 13 09:02:00 crc kubenswrapper[4685]: I1013 09:02:00.603199 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 13 09:02:00 crc kubenswrapper[4685]: I1013 09:02:00.868170 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 13 09:02:00 crc kubenswrapper[4685]: I1013 09:02:00.868493 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 13 09:02:01 crc kubenswrapper[4685]: I1013 09:02:01.889043 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 09:02:01 crc kubenswrapper[4685]: I1013 09:02:01.889079 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.491233 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-64vzp"] Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.499651 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-64vzp"] Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.589281 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.605566 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-64d69c974d-22pqc"] Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.605760 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-64d69c974d-22pqc" podUID="126e6e95-80d8-4e68-a541-8e53a4f5b862" containerName="horizon-log" containerID="cri-o://057d5f72232fe1033e30b206d78832bccc56b3093599e6d33773993e1467b3c1" gracePeriod=30 Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.607210 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-64d69c974d-22pqc" podUID="126e6e95-80d8-4e68-a541-8e53a4f5b862" containerName="horizon" containerID="cri-o://b115917cd44da7ffaccbe91d120ee61d59a4460f53f619a7e3e0215ca2922259" gracePeriod=30 Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.699027 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.730397 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-dbcf576b-msgft"] Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.731708 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.747438 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-dbcf576b-msgft"] Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.787531 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.827090 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-b29nc"] Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.829377 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b29nc" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.830426 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46272601-40bd-43ec-b7a8-6824281972e0-combined-ca-bundle\") pod \"horizon-dbcf576b-msgft\" (UID: \"46272601-40bd-43ec-b7a8-6824281972e0\") " pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.830484 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/46272601-40bd-43ec-b7a8-6824281972e0-horizon-secret-key\") pod \"horizon-dbcf576b-msgft\" (UID: \"46272601-40bd-43ec-b7a8-6824281972e0\") " pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.830553 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46272601-40bd-43ec-b7a8-6824281972e0-scripts\") pod \"horizon-dbcf576b-msgft\" (UID: \"46272601-40bd-43ec-b7a8-6824281972e0\") " pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.830725 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46272601-40bd-43ec-b7a8-6824281972e0-logs\") pod \"horizon-dbcf576b-msgft\" (UID: \"46272601-40bd-43ec-b7a8-6824281972e0\") " pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.830757 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/46272601-40bd-43ec-b7a8-6824281972e0-horizon-tls-certs\") pod \"horizon-dbcf576b-msgft\" (UID: \"46272601-40bd-43ec-b7a8-6824281972e0\") " pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.830813 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z67sd\" (UniqueName: \"kubernetes.io/projected/46272601-40bd-43ec-b7a8-6824281972e0-kube-api-access-z67sd\") pod \"horizon-dbcf576b-msgft\" (UID: \"46272601-40bd-43ec-b7a8-6824281972e0\") " pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.830838 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/46272601-40bd-43ec-b7a8-6824281972e0-config-data\") pod \"horizon-dbcf576b-msgft\" (UID: \"46272601-40bd-43ec-b7a8-6824281972e0\") " pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.834227 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.834278 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.834285 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.834386 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5xp9z" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.851810 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-b29nc"] Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.868012 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-55b747894d-xd2hx"] Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.869304 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.900779 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.900806 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.900956 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a9cfba62-3301-4ae3-b32c-d542302792d4" containerName="glance-log" containerID="cri-o://b645002af7b6e1599aca65be6d1ea012273493a144aa47de695155e939648e51" gracePeriod=30 Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.901061 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.901070 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.901137 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="af10022e-b9b5-4bed-8c80-9064b711f837" containerName="glance-log" containerID="cri-o://a5aee9e8730b63472aa83483317267b5ae11e7ad15304edd44922a53a600a0cb" gracePeriod=30 Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.901196 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a9cfba62-3301-4ae3-b32c-d542302792d4" containerName="glance-httpd" containerID="cri-o://6faa94ee30712a6d42db9f6e7ed17637643bdf663c088d057871184b5370d738" gracePeriod=30 Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.901443 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="af10022e-b9b5-4bed-8c80-9064b711f837" containerName="glance-httpd" containerID="cri-o://99516221fed8e294d62a6cb32d71ed431230ed3bc0d03203306b0083cec2be29" gracePeriod=30 Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.924189 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-55b747894d-xd2hx"] Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.932308 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a194ffb-9cf0-4167-9c5b-c51bd79c42d7-scripts\") pod \"placement-55b747894d-xd2hx\" (UID: \"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7\") " pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.932351 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46272601-40bd-43ec-b7a8-6824281972e0-scripts\") pod \"horizon-dbcf576b-msgft\" (UID: \"46272601-40bd-43ec-b7a8-6824281972e0\") " pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.932390 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tpb4\" (UniqueName: \"kubernetes.io/projected/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-kube-api-access-5tpb4\") pod \"keystone-bootstrap-b29nc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " pod="openstack/keystone-bootstrap-b29nc" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.932434 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-credential-keys\") pod \"keystone-bootstrap-b29nc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " pod="openstack/keystone-bootstrap-b29nc" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.932452 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-fernet-keys\") pod \"keystone-bootstrap-b29nc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " pod="openstack/keystone-bootstrap-b29nc" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.932473 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-config-data\") pod \"keystone-bootstrap-b29nc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " pod="openstack/keystone-bootstrap-b29nc" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.932496 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46272601-40bd-43ec-b7a8-6824281972e0-logs\") pod \"horizon-dbcf576b-msgft\" (UID: \"46272601-40bd-43ec-b7a8-6824281972e0\") " pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.932514 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-scripts\") pod \"keystone-bootstrap-b29nc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " pod="openstack/keystone-bootstrap-b29nc" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.932532 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a194ffb-9cf0-4167-9c5b-c51bd79c42d7-public-tls-certs\") pod \"placement-55b747894d-xd2hx\" (UID: \"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7\") " pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.932549 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/46272601-40bd-43ec-b7a8-6824281972e0-horizon-tls-certs\") pod \"horizon-dbcf576b-msgft\" (UID: \"46272601-40bd-43ec-b7a8-6824281972e0\") " pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.932581 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a194ffb-9cf0-4167-9c5b-c51bd79c42d7-internal-tls-certs\") pod \"placement-55b747894d-xd2hx\" (UID: \"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7\") " pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.932598 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vrst\" (UniqueName: \"kubernetes.io/projected/9a194ffb-9cf0-4167-9c5b-c51bd79c42d7-kube-api-access-9vrst\") pod \"placement-55b747894d-xd2hx\" (UID: \"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7\") " pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.932614 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z67sd\" (UniqueName: \"kubernetes.io/projected/46272601-40bd-43ec-b7a8-6824281972e0-kube-api-access-z67sd\") pod \"horizon-dbcf576b-msgft\" (UID: \"46272601-40bd-43ec-b7a8-6824281972e0\") " pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.932631 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/46272601-40bd-43ec-b7a8-6824281972e0-config-data\") pod \"horizon-dbcf576b-msgft\" (UID: \"46272601-40bd-43ec-b7a8-6824281972e0\") " pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.932654 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-combined-ca-bundle\") pod \"keystone-bootstrap-b29nc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " pod="openstack/keystone-bootstrap-b29nc" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.932680 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a194ffb-9cf0-4167-9c5b-c51bd79c42d7-config-data\") pod \"placement-55b747894d-xd2hx\" (UID: \"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7\") " pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.932702 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46272601-40bd-43ec-b7a8-6824281972e0-combined-ca-bundle\") pod \"horizon-dbcf576b-msgft\" (UID: \"46272601-40bd-43ec-b7a8-6824281972e0\") " pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.932698 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="a9cfba62-3301-4ae3-b32c-d542302792d4" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.149:9292/healthcheck\": EOF" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.932721 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a194ffb-9cf0-4167-9c5b-c51bd79c42d7-combined-ca-bundle\") pod \"placement-55b747894d-xd2hx\" (UID: \"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7\") " pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.932747 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/46272601-40bd-43ec-b7a8-6824281972e0-horizon-secret-key\") pod \"horizon-dbcf576b-msgft\" (UID: \"46272601-40bd-43ec-b7a8-6824281972e0\") " pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.932779 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a194ffb-9cf0-4167-9c5b-c51bd79c42d7-logs\") pod \"placement-55b747894d-xd2hx\" (UID: \"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7\") " pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.933431 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46272601-40bd-43ec-b7a8-6824281972e0-scripts\") pod \"horizon-dbcf576b-msgft\" (UID: \"46272601-40bd-43ec-b7a8-6824281972e0\") " pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.933627 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="a9cfba62-3301-4ae3-b32c-d542302792d4" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.149:9292/healthcheck\": EOF" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.933706 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46272601-40bd-43ec-b7a8-6824281972e0-logs\") pod \"horizon-dbcf576b-msgft\" (UID: \"46272601-40bd-43ec-b7a8-6824281972e0\") " pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.934827 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/46272601-40bd-43ec-b7a8-6824281972e0-config-data\") pod \"horizon-dbcf576b-msgft\" (UID: \"46272601-40bd-43ec-b7a8-6824281972e0\") " pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.940867 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46272601-40bd-43ec-b7a8-6824281972e0-combined-ca-bundle\") pod \"horizon-dbcf576b-msgft\" (UID: \"46272601-40bd-43ec-b7a8-6824281972e0\") " pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.943266 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/46272601-40bd-43ec-b7a8-6824281972e0-horizon-secret-key\") pod \"horizon-dbcf576b-msgft\" (UID: \"46272601-40bd-43ec-b7a8-6824281972e0\") " pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.982957 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="af10022e-b9b5-4bed-8c80-9064b711f837" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.150:9292/healthcheck\": EOF" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.983074 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="af10022e-b9b5-4bed-8c80-9064b711f837" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.150:9292/healthcheck\": EOF" Oct 13 09:02:02 crc kubenswrapper[4685]: I1013 09:02:02.998479 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z67sd\" (UniqueName: \"kubernetes.io/projected/46272601-40bd-43ec-b7a8-6824281972e0-kube-api-access-z67sd\") pod \"horizon-dbcf576b-msgft\" (UID: \"46272601-40bd-43ec-b7a8-6824281972e0\") " pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.010416 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/46272601-40bd-43ec-b7a8-6824281972e0-horizon-tls-certs\") pod \"horizon-dbcf576b-msgft\" (UID: \"46272601-40bd-43ec-b7a8-6824281972e0\") " pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.034768 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-scripts\") pod \"keystone-bootstrap-b29nc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " pod="openstack/keystone-bootstrap-b29nc" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.034812 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a194ffb-9cf0-4167-9c5b-c51bd79c42d7-public-tls-certs\") pod \"placement-55b747894d-xd2hx\" (UID: \"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7\") " pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.034848 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a194ffb-9cf0-4167-9c5b-c51bd79c42d7-internal-tls-certs\") pod \"placement-55b747894d-xd2hx\" (UID: \"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7\") " pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.034865 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vrst\" (UniqueName: \"kubernetes.io/projected/9a194ffb-9cf0-4167-9c5b-c51bd79c42d7-kube-api-access-9vrst\") pod \"placement-55b747894d-xd2hx\" (UID: \"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7\") " pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.034885 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-combined-ca-bundle\") pod \"keystone-bootstrap-b29nc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " pod="openstack/keystone-bootstrap-b29nc" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.034927 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a194ffb-9cf0-4167-9c5b-c51bd79c42d7-config-data\") pod \"placement-55b747894d-xd2hx\" (UID: \"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7\") " pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.034949 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a194ffb-9cf0-4167-9c5b-c51bd79c42d7-combined-ca-bundle\") pod \"placement-55b747894d-xd2hx\" (UID: \"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7\") " pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.034978 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a194ffb-9cf0-4167-9c5b-c51bd79c42d7-logs\") pod \"placement-55b747894d-xd2hx\" (UID: \"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7\") " pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.034996 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a194ffb-9cf0-4167-9c5b-c51bd79c42d7-scripts\") pod \"placement-55b747894d-xd2hx\" (UID: \"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7\") " pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.035029 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tpb4\" (UniqueName: \"kubernetes.io/projected/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-kube-api-access-5tpb4\") pod \"keystone-bootstrap-b29nc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " pod="openstack/keystone-bootstrap-b29nc" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.035066 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-credential-keys\") pod \"keystone-bootstrap-b29nc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " pod="openstack/keystone-bootstrap-b29nc" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.035083 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-fernet-keys\") pod \"keystone-bootstrap-b29nc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " pod="openstack/keystone-bootstrap-b29nc" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.035106 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-config-data\") pod \"keystone-bootstrap-b29nc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " pod="openstack/keystone-bootstrap-b29nc" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.042719 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-config-data\") pod \"keystone-bootstrap-b29nc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " pod="openstack/keystone-bootstrap-b29nc" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.042837 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a194ffb-9cf0-4167-9c5b-c51bd79c42d7-internal-tls-certs\") pod \"placement-55b747894d-xd2hx\" (UID: \"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7\") " pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.046357 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a194ffb-9cf0-4167-9c5b-c51bd79c42d7-logs\") pod \"placement-55b747894d-xd2hx\" (UID: \"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7\") " pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.046853 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a194ffb-9cf0-4167-9c5b-c51bd79c42d7-config-data\") pod \"placement-55b747894d-xd2hx\" (UID: \"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7\") " pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.047778 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.059617 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-fernet-keys\") pod \"keystone-bootstrap-b29nc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " pod="openstack/keystone-bootstrap-b29nc" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.060390 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a194ffb-9cf0-4167-9c5b-c51bd79c42d7-scripts\") pod \"placement-55b747894d-xd2hx\" (UID: \"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7\") " pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.060615 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-scripts\") pod \"keystone-bootstrap-b29nc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " pod="openstack/keystone-bootstrap-b29nc" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.061334 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a194ffb-9cf0-4167-9c5b-c51bd79c42d7-public-tls-certs\") pod \"placement-55b747894d-xd2hx\" (UID: \"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7\") " pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.061926 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a194ffb-9cf0-4167-9c5b-c51bd79c42d7-combined-ca-bundle\") pod \"placement-55b747894d-xd2hx\" (UID: \"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7\") " pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.064646 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vrst\" (UniqueName: \"kubernetes.io/projected/9a194ffb-9cf0-4167-9c5b-c51bd79c42d7-kube-api-access-9vrst\") pod \"placement-55b747894d-xd2hx\" (UID: \"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7\") " pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.065778 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tpb4\" (UniqueName: \"kubernetes.io/projected/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-kube-api-access-5tpb4\") pod \"keystone-bootstrap-b29nc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " pod="openstack/keystone-bootstrap-b29nc" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.075402 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-credential-keys\") pod \"keystone-bootstrap-b29nc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " pod="openstack/keystone-bootstrap-b29nc" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.076441 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-combined-ca-bundle\") pod \"keystone-bootstrap-b29nc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " pod="openstack/keystone-bootstrap-b29nc" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.144187 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b29nc" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.203976 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.511507 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b23db04-161e-4736-85de-a2173dfa0d4f" path="/var/lib/kubelet/pods/8b23db04-161e-4736-85de-a2173dfa0d4f/volumes" Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.910188 4685 generic.go:334] "Generic (PLEG): container finished" podID="a9cfba62-3301-4ae3-b32c-d542302792d4" containerID="b645002af7b6e1599aca65be6d1ea012273493a144aa47de695155e939648e51" exitCode=143 Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.910243 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a9cfba62-3301-4ae3-b32c-d542302792d4","Type":"ContainerDied","Data":"b645002af7b6e1599aca65be6d1ea012273493a144aa47de695155e939648e51"} Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.911801 4685 generic.go:334] "Generic (PLEG): container finished" podID="af10022e-b9b5-4bed-8c80-9064b711f837" containerID="a5aee9e8730b63472aa83483317267b5ae11e7ad15304edd44922a53a600a0cb" exitCode=143 Oct 13 09:02:03 crc kubenswrapper[4685]: I1013 09:02:03.911825 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"af10022e-b9b5-4bed-8c80-9064b711f837","Type":"ContainerDied","Data":"a5aee9e8730b63472aa83483317267b5ae11e7ad15304edd44922a53a600a0cb"} Oct 13 09:02:05 crc kubenswrapper[4685]: I1013 09:02:05.525360 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 13 09:02:05 crc kubenswrapper[4685]: I1013 09:02:05.528347 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 13 09:02:08 crc kubenswrapper[4685]: I1013 09:02:08.364249 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="af10022e-b9b5-4bed-8c80-9064b711f837" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.150:9292/healthcheck\": read tcp 10.217.0.2:55460->10.217.0.150:9292: read: connection reset by peer" Oct 13 09:02:08 crc kubenswrapper[4685]: I1013 09:02:08.365710 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="af10022e-b9b5-4bed-8c80-9064b711f837" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.150:9292/healthcheck\": read tcp 10.217.0.2:55456->10.217.0.150:9292: read: connection reset by peer" Oct 13 09:02:08 crc kubenswrapper[4685]: I1013 09:02:08.973707 4685 generic.go:334] "Generic (PLEG): container finished" podID="af10022e-b9b5-4bed-8c80-9064b711f837" containerID="99516221fed8e294d62a6cb32d71ed431230ed3bc0d03203306b0083cec2be29" exitCode=0 Oct 13 09:02:08 crc kubenswrapper[4685]: I1013 09:02:08.973773 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"af10022e-b9b5-4bed-8c80-9064b711f837","Type":"ContainerDied","Data":"99516221fed8e294d62a6cb32d71ed431230ed3bc0d03203306b0083cec2be29"} Oct 13 09:02:09 crc kubenswrapper[4685]: I1013 09:02:09.269023 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-db976b79d-q446d" podUID="2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 13 09:02:09 crc kubenswrapper[4685]: I1013 09:02:09.993070 4685 generic.go:334] "Generic (PLEG): container finished" podID="a9cfba62-3301-4ae3-b32c-d542302792d4" containerID="6faa94ee30712a6d42db9f6e7ed17637643bdf663c088d057871184b5370d738" exitCode=0 Oct 13 09:02:09 crc kubenswrapper[4685]: I1013 09:02:09.993134 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a9cfba62-3301-4ae3-b32c-d542302792d4","Type":"ContainerDied","Data":"6faa94ee30712a6d42db9f6e7ed17637643bdf663c088d057871184b5370d738"} Oct 13 09:02:13 crc kubenswrapper[4685]: I1013 09:02:13.031789 4685 generic.go:334] "Generic (PLEG): container finished" podID="ee75d00f-3165-45a5-854d-1045b43f2c87" containerID="ee542b97c2ea580a1848390372ad4dd4d87e49fcb174c14691b7a9017f8a1a5c" exitCode=0 Oct 13 09:02:13 crc kubenswrapper[4685]: I1013 09:02:13.032102 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9msgq" event={"ID":"ee75d00f-3165-45a5-854d-1045b43f2c87","Type":"ContainerDied","Data":"ee542b97c2ea580a1848390372ad4dd4d87e49fcb174c14691b7a9017f8a1a5c"} Oct 13 09:02:15 crc kubenswrapper[4685]: E1013 09:02:15.019001 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 13 09:02:15 crc kubenswrapper[4685]: E1013 09:02:15.019707 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fmq9p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-7h6qb_openstack(b225d240-40a0-4d59-aa19-6355515e9f8c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 09:02:15 crc kubenswrapper[4685]: E1013 09:02:15.020839 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-7h6qb" podUID="b225d240-40a0-4d59-aa19-6355515e9f8c" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.056901 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a9cfba62-3301-4ae3-b32c-d542302792d4","Type":"ContainerDied","Data":"fa655b2dfe5494ba5b83f8c375a3bb91ead32f8e8f09769d1ec1ff604aaa9ac1"} Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.056971 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa655b2dfe5494ba5b83f8c375a3bb91ead32f8e8f09769d1ec1ff604aaa9ac1" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.060640 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"af10022e-b9b5-4bed-8c80-9064b711f837","Type":"ContainerDied","Data":"08fe3b449e8aeacd341757b0b6b87e803db87305a67a377cc3bf7c3c04ee6e11"} Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.060685 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08fe3b449e8aeacd341757b0b6b87e803db87305a67a377cc3bf7c3c04ee6e11" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.063973 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9msgq" event={"ID":"ee75d00f-3165-45a5-854d-1045b43f2c87","Type":"ContainerDied","Data":"d9f5d6878800827b2e5dcbd3db3436e3ac03c905078486a523d045f2ed50a2eb"} Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.064029 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9f5d6878800827b2e5dcbd3db3436e3ac03c905078486a523d045f2ed50a2eb" Oct 13 09:02:15 crc kubenswrapper[4685]: E1013 09:02:15.072201 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-7h6qb" podUID="b225d240-40a0-4d59-aa19-6355515e9f8c" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.261805 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.285899 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9msgq" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.311280 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.386194 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-scripts\") pod \"af10022e-b9b5-4bed-8c80-9064b711f837\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.387195 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hpdn\" (UniqueName: \"kubernetes.io/projected/af10022e-b9b5-4bed-8c80-9064b711f837-kube-api-access-5hpdn\") pod \"af10022e-b9b5-4bed-8c80-9064b711f837\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.387351 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af10022e-b9b5-4bed-8c80-9064b711f837-logs\") pod \"af10022e-b9b5-4bed-8c80-9064b711f837\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.387498 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/af10022e-b9b5-4bed-8c80-9064b711f837-httpd-run\") pod \"af10022e-b9b5-4bed-8c80-9064b711f837\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.387616 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-combined-ca-bundle\") pod \"af10022e-b9b5-4bed-8c80-9064b711f837\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.387715 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee75d00f-3165-45a5-854d-1045b43f2c87-combined-ca-bundle\") pod \"ee75d00f-3165-45a5-854d-1045b43f2c87\" (UID: \"ee75d00f-3165-45a5-854d-1045b43f2c87\") " Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.387777 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"af10022e-b9b5-4bed-8c80-9064b711f837\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.387871 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-config-data\") pod \"af10022e-b9b5-4bed-8c80-9064b711f837\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.387949 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af10022e-b9b5-4bed-8c80-9064b711f837-logs" (OuterVolumeSpecName: "logs") pod "af10022e-b9b5-4bed-8c80-9064b711f837" (UID: "af10022e-b9b5-4bed-8c80-9064b711f837"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.388051 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ee75d00f-3165-45a5-854d-1045b43f2c87-config\") pod \"ee75d00f-3165-45a5-854d-1045b43f2c87\" (UID: \"ee75d00f-3165-45a5-854d-1045b43f2c87\") " Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.388293 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-internal-tls-certs\") pod \"af10022e-b9b5-4bed-8c80-9064b711f837\" (UID: \"af10022e-b9b5-4bed-8c80-9064b711f837\") " Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.388373 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rppfh\" (UniqueName: \"kubernetes.io/projected/ee75d00f-3165-45a5-854d-1045b43f2c87-kube-api-access-rppfh\") pod \"ee75d00f-3165-45a5-854d-1045b43f2c87\" (UID: \"ee75d00f-3165-45a5-854d-1045b43f2c87\") " Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.388818 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af10022e-b9b5-4bed-8c80-9064b711f837-logs\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.399314 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-scripts" (OuterVolumeSpecName: "scripts") pod "af10022e-b9b5-4bed-8c80-9064b711f837" (UID: "af10022e-b9b5-4bed-8c80-9064b711f837"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.399695 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af10022e-b9b5-4bed-8c80-9064b711f837-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "af10022e-b9b5-4bed-8c80-9064b711f837" (UID: "af10022e-b9b5-4bed-8c80-9064b711f837"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.411439 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee75d00f-3165-45a5-854d-1045b43f2c87-kube-api-access-rppfh" (OuterVolumeSpecName: "kube-api-access-rppfh") pod "ee75d00f-3165-45a5-854d-1045b43f2c87" (UID: "ee75d00f-3165-45a5-854d-1045b43f2c87"). InnerVolumeSpecName "kube-api-access-rppfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.411557 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "af10022e-b9b5-4bed-8c80-9064b711f837" (UID: "af10022e-b9b5-4bed-8c80-9064b711f837"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.411624 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af10022e-b9b5-4bed-8c80-9064b711f837-kube-api-access-5hpdn" (OuterVolumeSpecName: "kube-api-access-5hpdn") pod "af10022e-b9b5-4bed-8c80-9064b711f837" (UID: "af10022e-b9b5-4bed-8c80-9064b711f837"). InnerVolumeSpecName "kube-api-access-5hpdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.443480 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee75d00f-3165-45a5-854d-1045b43f2c87-config" (OuterVolumeSpecName: "config") pod "ee75d00f-3165-45a5-854d-1045b43f2c87" (UID: "ee75d00f-3165-45a5-854d-1045b43f2c87"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.456324 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af10022e-b9b5-4bed-8c80-9064b711f837" (UID: "af10022e-b9b5-4bed-8c80-9064b711f837"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.481380 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee75d00f-3165-45a5-854d-1045b43f2c87-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee75d00f-3165-45a5-854d-1045b43f2c87" (UID: "ee75d00f-3165-45a5-854d-1045b43f2c87"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.489551 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"a9cfba62-3301-4ae3-b32c-d542302792d4\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.489609 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-combined-ca-bundle\") pod \"a9cfba62-3301-4ae3-b32c-d542302792d4\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.489696 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a9cfba62-3301-4ae3-b32c-d542302792d4-httpd-run\") pod \"a9cfba62-3301-4ae3-b32c-d542302792d4\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.489730 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9cfba62-3301-4ae3-b32c-d542302792d4-logs\") pod \"a9cfba62-3301-4ae3-b32c-d542302792d4\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.489781 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-config-data\") pod \"a9cfba62-3301-4ae3-b32c-d542302792d4\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.489808 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-public-tls-certs\") pod \"a9cfba62-3301-4ae3-b32c-d542302792d4\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.489876 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xh78\" (UniqueName: \"kubernetes.io/projected/a9cfba62-3301-4ae3-b32c-d542302792d4-kube-api-access-5xh78\") pod \"a9cfba62-3301-4ae3-b32c-d542302792d4\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.489894 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-scripts\") pod \"a9cfba62-3301-4ae3-b32c-d542302792d4\" (UID: \"a9cfba62-3301-4ae3-b32c-d542302792d4\") " Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.490242 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/ee75d00f-3165-45a5-854d-1045b43f2c87-config\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.490254 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rppfh\" (UniqueName: \"kubernetes.io/projected/ee75d00f-3165-45a5-854d-1045b43f2c87-kube-api-access-rppfh\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.490263 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.490271 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hpdn\" (UniqueName: \"kubernetes.io/projected/af10022e-b9b5-4bed-8c80-9064b711f837-kube-api-access-5hpdn\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.490279 4685 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/af10022e-b9b5-4bed-8c80-9064b711f837-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.490288 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.490298 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee75d00f-3165-45a5-854d-1045b43f2c87-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.490316 4685 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.492142 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9cfba62-3301-4ae3-b32c-d542302792d4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a9cfba62-3301-4ae3-b32c-d542302792d4" (UID: "a9cfba62-3301-4ae3-b32c-d542302792d4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.494724 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9cfba62-3301-4ae3-b32c-d542302792d4-logs" (OuterVolumeSpecName: "logs") pod "a9cfba62-3301-4ae3-b32c-d542302792d4" (UID: "a9cfba62-3301-4ae3-b32c-d542302792d4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.497243 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "af10022e-b9b5-4bed-8c80-9064b711f837" (UID: "af10022e-b9b5-4bed-8c80-9064b711f837"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.498029 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9cfba62-3301-4ae3-b32c-d542302792d4-kube-api-access-5xh78" (OuterVolumeSpecName: "kube-api-access-5xh78") pod "a9cfba62-3301-4ae3-b32c-d542302792d4" (UID: "a9cfba62-3301-4ae3-b32c-d542302792d4"). InnerVolumeSpecName "kube-api-access-5xh78". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.502500 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-scripts" (OuterVolumeSpecName: "scripts") pod "a9cfba62-3301-4ae3-b32c-d542302792d4" (UID: "a9cfba62-3301-4ae3-b32c-d542302792d4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.523160 4685 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.531544 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "a9cfba62-3301-4ae3-b32c-d542302792d4" (UID: "a9cfba62-3301-4ae3-b32c-d542302792d4"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.557360 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-config-data" (OuterVolumeSpecName: "config-data") pod "af10022e-b9b5-4bed-8c80-9064b711f837" (UID: "af10022e-b9b5-4bed-8c80-9064b711f837"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.565438 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-config-data" (OuterVolumeSpecName: "config-data") pod "a9cfba62-3301-4ae3-b32c-d542302792d4" (UID: "a9cfba62-3301-4ae3-b32c-d542302792d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.574906 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9cfba62-3301-4ae3-b32c-d542302792d4" (UID: "a9cfba62-3301-4ae3-b32c-d542302792d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.594431 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xh78\" (UniqueName: \"kubernetes.io/projected/a9cfba62-3301-4ae3-b32c-d542302792d4-kube-api-access-5xh78\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.594491 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.594561 4685 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.594577 4685 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.594590 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.594602 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.594638 4685 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/af10022e-b9b5-4bed-8c80-9064b711f837-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.594652 4685 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a9cfba62-3301-4ae3-b32c-d542302792d4-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.594662 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9cfba62-3301-4ae3-b32c-d542302792d4-logs\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.594673 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.614631 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7d4ff599b8-7thr6"] Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.619346 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a9cfba62-3301-4ae3-b32c-d542302792d4" (UID: "a9cfba62-3301-4ae3-b32c-d542302792d4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.625028 4685 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 13 09:02:15 crc kubenswrapper[4685]: W1013 09:02:15.670594 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40fbc731_ab15_4cc0_9eab_1ec6c0848cbc.slice/crio-4b103562677017cda82e8fbbc57462022313bf928ac9f7711e1daaba740458ac WatchSource:0}: Error finding container 4b103562677017cda82e8fbbc57462022313bf928ac9f7711e1daaba740458ac: Status 404 returned error can't find the container with id 4b103562677017cda82e8fbbc57462022313bf928ac9f7711e1daaba740458ac Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.674327 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-b29nc"] Oct 13 09:02:15 crc kubenswrapper[4685]: W1013 09:02:15.697131 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46272601_40bd_43ec_b7a8_6824281972e0.slice/crio-205542f17a90de9fb210ef96e95ac41e1486c8f9f380aaf49237e9b14069501b WatchSource:0}: Error finding container 205542f17a90de9fb210ef96e95ac41e1486c8f9f380aaf49237e9b14069501b: Status 404 returned error can't find the container with id 205542f17a90de9fb210ef96e95ac41e1486c8f9f380aaf49237e9b14069501b Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.698366 4685 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9cfba62-3301-4ae3-b32c-d542302792d4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.698446 4685 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.703122 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-dbcf576b-msgft"] Oct 13 09:02:15 crc kubenswrapper[4685]: I1013 09:02:15.789451 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-55b747894d-xd2hx"] Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.078839 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-55b747894d-xd2hx" event={"ID":"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7","Type":"ContainerStarted","Data":"f864bc37c79f1cbaf404cbd4e14dc2ac8219673e7e41895c5e31052a0444c69a"} Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.079227 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-55b747894d-xd2hx" event={"ID":"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7","Type":"ContainerStarted","Data":"f3b0575976842d0f2cef9633e36410961be87b2c4dc6ac9c85d0a73469ccfb99"} Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.082406 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d4ff599b8-7thr6" event={"ID":"19f3def1-e242-41ec-8465-164d53d878a4","Type":"ContainerStarted","Data":"5ed93862142f8804adadb512e47f39fd79361493df4ab86de88674ec6691104b"} Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.082464 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d4ff599b8-7thr6" event={"ID":"19f3def1-e242-41ec-8465-164d53d878a4","Type":"ContainerStarted","Data":"2ceb06eecac6b5172b62220dc937acd208d280a46058c9d9b0cb2269ce2785ed"} Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.083849 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5p6fr" event={"ID":"05506d41-5449-4eb6-86ce-0ac5f5c880ba","Type":"ContainerStarted","Data":"114f444907b28763ac128396f5487639845624f6c74051d43cbdffb59c484704"} Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.091138 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-dbcf576b-msgft" event={"ID":"46272601-40bd-43ec-b7a8-6824281972e0","Type":"ContainerStarted","Data":"0c438886491b1cdfc7428bfe0d29d0a9723284eb295eb41c3130df40bef669d0"} Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.091178 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-dbcf576b-msgft" event={"ID":"46272601-40bd-43ec-b7a8-6824281972e0","Type":"ContainerStarted","Data":"205542f17a90de9fb210ef96e95ac41e1486c8f9f380aaf49237e9b14069501b"} Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.094497 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b29nc" event={"ID":"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc","Type":"ContainerStarted","Data":"16c5d0f261753dbe0cc387ffa98a76a6a4a86536f143165e934db71b95f53192"} Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.094545 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b29nc" event={"ID":"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc","Type":"ContainerStarted","Data":"4b103562677017cda82e8fbbc57462022313bf928ac9f7711e1daaba740458ac"} Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.108727 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.108958 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-5p6fr" podStartSLOduration=14.419832956 podStartE2EDuration="40.108941554s" podCreationTimestamp="2025-10-13 09:01:36 +0000 UTC" firstStartedPulling="2025-10-13 09:01:49.293048569 +0000 UTC m=+1034.440924330" lastFinishedPulling="2025-10-13 09:02:14.982157147 +0000 UTC m=+1060.130032928" observedRunningTime="2025-10-13 09:02:16.100169464 +0000 UTC m=+1061.248045235" watchObservedRunningTime="2025-10-13 09:02:16.108941554 +0000 UTC m=+1061.256817315" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.109026 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ffc408c-86d3-4e93-9d49-6ef39cb7d959","Type":"ContainerStarted","Data":"dfd28dc5a0e2852cf335562850401d5db7a9d724c0154c5edec7099ac902d449"} Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.109860 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.109890 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9msgq" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.135095 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-b29nc" podStartSLOduration=14.135076057 podStartE2EDuration="14.135076057s" podCreationTimestamp="2025-10-13 09:02:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:02:16.12380952 +0000 UTC m=+1061.271685281" watchObservedRunningTime="2025-10-13 09:02:16.135076057 +0000 UTC m=+1061.282951818" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.284806 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.296336 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.337001 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 09:02:16 crc kubenswrapper[4685]: E1013 09:02:16.337433 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9cfba62-3301-4ae3-b32c-d542302792d4" containerName="glance-log" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.337451 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9cfba62-3301-4ae3-b32c-d542302792d4" containerName="glance-log" Oct 13 09:02:16 crc kubenswrapper[4685]: E1013 09:02:16.337468 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af10022e-b9b5-4bed-8c80-9064b711f837" containerName="glance-log" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.337475 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="af10022e-b9b5-4bed-8c80-9064b711f837" containerName="glance-log" Oct 13 09:02:16 crc kubenswrapper[4685]: E1013 09:02:16.337499 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af10022e-b9b5-4bed-8c80-9064b711f837" containerName="glance-httpd" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.337505 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="af10022e-b9b5-4bed-8c80-9064b711f837" containerName="glance-httpd" Oct 13 09:02:16 crc kubenswrapper[4685]: E1013 09:02:16.337520 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee75d00f-3165-45a5-854d-1045b43f2c87" containerName="neutron-db-sync" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.337526 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee75d00f-3165-45a5-854d-1045b43f2c87" containerName="neutron-db-sync" Oct 13 09:02:16 crc kubenswrapper[4685]: E1013 09:02:16.337544 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9cfba62-3301-4ae3-b32c-d542302792d4" containerName="glance-httpd" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.337550 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9cfba62-3301-4ae3-b32c-d542302792d4" containerName="glance-httpd" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.337741 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="af10022e-b9b5-4bed-8c80-9064b711f837" containerName="glance-log" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.337753 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9cfba62-3301-4ae3-b32c-d542302792d4" containerName="glance-log" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.337771 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="af10022e-b9b5-4bed-8c80-9064b711f837" containerName="glance-httpd" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.337781 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee75d00f-3165-45a5-854d-1045b43f2c87" containerName="neutron-db-sync" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.337790 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9cfba62-3301-4ae3-b32c-d542302792d4" containerName="glance-httpd" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.340074 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.342692 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.343091 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gt2pf" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.343265 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.345142 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.369263 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.398429 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.416881 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.422405 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.422462 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-scripts\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.422504 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-config-data\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.422524 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.422545 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e81aae5c-c9f2-455c-8154-fa64d948bcab-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.422626 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e81aae5c-c9f2-455c-8154-fa64d948bcab-logs\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.422644 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n2jc\" (UniqueName: \"kubernetes.io/projected/e81aae5c-c9f2-455c-8154-fa64d948bcab-kube-api-access-9n2jc\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.422661 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.422977 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.424510 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.434140 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.434428 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.436298 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.523809 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-scripts\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.523864 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.523887 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.523927 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-config-data\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.523954 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.523980 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/babeefc6-9cbe-4f6c-9487-731a9270c1c4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.524003 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e81aae5c-c9f2-455c-8154-fa64d948bcab-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.524070 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.524097 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.524118 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/babeefc6-9cbe-4f6c-9487-731a9270c1c4-logs\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.524135 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e81aae5c-c9f2-455c-8154-fa64d948bcab-logs\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.524156 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n2jc\" (UniqueName: \"kubernetes.io/projected/e81aae5c-c9f2-455c-8154-fa64d948bcab-kube-api-access-9n2jc\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.524174 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.524194 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.524210 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.524228 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpnbw\" (UniqueName: \"kubernetes.io/projected/babeefc6-9cbe-4f6c-9487-731a9270c1c4-kube-api-access-tpnbw\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.535726 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e81aae5c-c9f2-455c-8154-fa64d948bcab-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.535964 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e81aae5c-c9f2-455c-8154-fa64d948bcab-logs\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.536138 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.539884 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-scripts\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.545854 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.547678 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.552039 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-config-data\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.569727 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n2jc\" (UniqueName: \"kubernetes.io/projected/e81aae5c-c9f2-455c-8154-fa64d948bcab-kube-api-access-9n2jc\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.625757 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.625797 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/babeefc6-9cbe-4f6c-9487-731a9270c1c4-logs\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.625851 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.625866 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpnbw\" (UniqueName: \"kubernetes.io/projected/babeefc6-9cbe-4f6c-9487-731a9270c1c4-kube-api-access-tpnbw\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.625906 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.625938 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.625974 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/babeefc6-9cbe-4f6c-9487-731a9270c1c4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.626051 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.626576 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.627000 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/babeefc6-9cbe-4f6c-9487-731a9270c1c4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.630136 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/babeefc6-9cbe-4f6c-9487-731a9270c1c4-logs\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.634861 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.641408 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.685028 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.685443 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.695627 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpnbw\" (UniqueName: \"kubernetes.io/projected/babeefc6-9cbe-4f6c-9487-731a9270c1c4-kube-api-access-tpnbw\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.725461 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.729483 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.779418 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.819614 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-5rqpr"] Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.847298 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.865300 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-5rqpr"] Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.930540 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-dns-svc\") pod \"dnsmasq-dns-6b7b667979-5rqpr\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.930594 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-5rqpr\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.930618 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-5rqpr\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.930639 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bp67s\" (UniqueName: \"kubernetes.io/projected/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-kube-api-access-bp67s\") pod \"dnsmasq-dns-6b7b667979-5rqpr\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.930664 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-5rqpr\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.930748 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-config\") pod \"dnsmasq-dns-6b7b667979-5rqpr\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.967807 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.973006 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5c8fd5c86b-qkv7x"] Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.974446 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c8fd5c86b-qkv7x" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.977956 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.978183 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-6hrl6" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.979833 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 13 09:02:16 crc kubenswrapper[4685]: I1013 09:02:16.985599 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.039988 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5c8fd5c86b-qkv7x"] Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.040213 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-combined-ca-bundle\") pod \"neutron-5c8fd5c86b-qkv7x\" (UID: \"89e018d5-d463-4b0e-8fef-f1d3230d703c\") " pod="openstack/neutron-5c8fd5c86b-qkv7x" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.040273 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-ovndb-tls-certs\") pod \"neutron-5c8fd5c86b-qkv7x\" (UID: \"89e018d5-d463-4b0e-8fef-f1d3230d703c\") " pod="openstack/neutron-5c8fd5c86b-qkv7x" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.040304 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-config\") pod \"dnsmasq-dns-6b7b667979-5rqpr\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.040342 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g426\" (UniqueName: \"kubernetes.io/projected/89e018d5-d463-4b0e-8fef-f1d3230d703c-kube-api-access-6g426\") pod \"neutron-5c8fd5c86b-qkv7x\" (UID: \"89e018d5-d463-4b0e-8fef-f1d3230d703c\") " pod="openstack/neutron-5c8fd5c86b-qkv7x" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.040365 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-config\") pod \"neutron-5c8fd5c86b-qkv7x\" (UID: \"89e018d5-d463-4b0e-8fef-f1d3230d703c\") " pod="openstack/neutron-5c8fd5c86b-qkv7x" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.040386 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-dns-svc\") pod \"dnsmasq-dns-6b7b667979-5rqpr\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.040408 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-5rqpr\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.040435 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-httpd-config\") pod \"neutron-5c8fd5c86b-qkv7x\" (UID: \"89e018d5-d463-4b0e-8fef-f1d3230d703c\") " pod="openstack/neutron-5c8fd5c86b-qkv7x" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.040455 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-5rqpr\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.040475 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bp67s\" (UniqueName: \"kubernetes.io/projected/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-kube-api-access-bp67s\") pod \"dnsmasq-dns-6b7b667979-5rqpr\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.040497 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-5rqpr\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.041812 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-config\") pod \"dnsmasq-dns-6b7b667979-5rqpr\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.041837 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-dns-svc\") pod \"dnsmasq-dns-6b7b667979-5rqpr\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.042456 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-5rqpr\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.042503 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-5rqpr\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.059282 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-5rqpr\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.080675 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bp67s\" (UniqueName: \"kubernetes.io/projected/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-kube-api-access-bp67s\") pod \"dnsmasq-dns-6b7b667979-5rqpr\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.142819 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-combined-ca-bundle\") pod \"neutron-5c8fd5c86b-qkv7x\" (UID: \"89e018d5-d463-4b0e-8fef-f1d3230d703c\") " pod="openstack/neutron-5c8fd5c86b-qkv7x" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.142869 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-ovndb-tls-certs\") pod \"neutron-5c8fd5c86b-qkv7x\" (UID: \"89e018d5-d463-4b0e-8fef-f1d3230d703c\") " pod="openstack/neutron-5c8fd5c86b-qkv7x" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.142925 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g426\" (UniqueName: \"kubernetes.io/projected/89e018d5-d463-4b0e-8fef-f1d3230d703c-kube-api-access-6g426\") pod \"neutron-5c8fd5c86b-qkv7x\" (UID: \"89e018d5-d463-4b0e-8fef-f1d3230d703c\") " pod="openstack/neutron-5c8fd5c86b-qkv7x" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.142948 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-config\") pod \"neutron-5c8fd5c86b-qkv7x\" (UID: \"89e018d5-d463-4b0e-8fef-f1d3230d703c\") " pod="openstack/neutron-5c8fd5c86b-qkv7x" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.142979 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-httpd-config\") pod \"neutron-5c8fd5c86b-qkv7x\" (UID: \"89e018d5-d463-4b0e-8fef-f1d3230d703c\") " pod="openstack/neutron-5c8fd5c86b-qkv7x" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.169647 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-httpd-config\") pod \"neutron-5c8fd5c86b-qkv7x\" (UID: \"89e018d5-d463-4b0e-8fef-f1d3230d703c\") " pod="openstack/neutron-5c8fd5c86b-qkv7x" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.169983 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-config\") pod \"neutron-5c8fd5c86b-qkv7x\" (UID: \"89e018d5-d463-4b0e-8fef-f1d3230d703c\") " pod="openstack/neutron-5c8fd5c86b-qkv7x" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.170964 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-ovndb-tls-certs\") pod \"neutron-5c8fd5c86b-qkv7x\" (UID: \"89e018d5-d463-4b0e-8fef-f1d3230d703c\") " pod="openstack/neutron-5c8fd5c86b-qkv7x" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.179520 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.180315 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-combined-ca-bundle\") pod \"neutron-5c8fd5c86b-qkv7x\" (UID: \"89e018d5-d463-4b0e-8fef-f1d3230d703c\") " pod="openstack/neutron-5c8fd5c86b-qkv7x" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.183327 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-55b747894d-xd2hx" event={"ID":"9a194ffb-9cf0-4167-9c5b-c51bd79c42d7","Type":"ContainerStarted","Data":"8ef5e9e5f2246260414d9a37d387f38cbab6b118f5b4398ac8e01082b595f758"} Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.184593 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.184617 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.202954 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d4ff599b8-7thr6" event={"ID":"19f3def1-e242-41ec-8465-164d53d878a4","Type":"ContainerStarted","Data":"3a44a77bb33a49bbaad3ced919a264013171d40d6bba7c0e36f5838e6c2ba239"} Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.203703 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.203724 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.231859 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g426\" (UniqueName: \"kubernetes.io/projected/89e018d5-d463-4b0e-8fef-f1d3230d703c-kube-api-access-6g426\") pod \"neutron-5c8fd5c86b-qkv7x\" (UID: \"89e018d5-d463-4b0e-8fef-f1d3230d703c\") " pod="openstack/neutron-5c8fd5c86b-qkv7x" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.255272 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-dbcf576b-msgft" event={"ID":"46272601-40bd-43ec-b7a8-6824281972e0","Type":"ContainerStarted","Data":"086ee9b5bb7db275108b5dea7888c4325943b8f930d8f0bedf8647a954bc2b2f"} Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.324171 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-55b747894d-xd2hx" podStartSLOduration=15.324146623 podStartE2EDuration="15.324146623s" podCreationTimestamp="2025-10-13 09:02:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:02:17.229355016 +0000 UTC m=+1062.377230787" watchObservedRunningTime="2025-10-13 09:02:17.324146623 +0000 UTC m=+1062.472022384" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.358616 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7d4ff599b8-7thr6" podStartSLOduration=19.358600195 podStartE2EDuration="19.358600195s" podCreationTimestamp="2025-10-13 09:01:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:02:17.285648563 +0000 UTC m=+1062.433524324" watchObservedRunningTime="2025-10-13 09:02:17.358600195 +0000 UTC m=+1062.506475956" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.417327 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-dbcf576b-msgft" podStartSLOduration=15.417303487 podStartE2EDuration="15.417303487s" podCreationTimestamp="2025-10-13 09:02:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:02:17.385064307 +0000 UTC m=+1062.532940068" watchObservedRunningTime="2025-10-13 09:02:17.417303487 +0000 UTC m=+1062.565179248" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.453513 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c8fd5c86b-qkv7x" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.531781 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9cfba62-3301-4ae3-b32c-d542302792d4" path="/var/lib/kubelet/pods/a9cfba62-3301-4ae3-b32c-d542302792d4/volumes" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.532782 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af10022e-b9b5-4bed-8c80-9064b711f837" path="/var/lib/kubelet/pods/af10022e-b9b5-4bed-8c80-9064b711f837/volumes" Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.675168 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.855006 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 09:02:17 crc kubenswrapper[4685]: I1013 09:02:17.901635 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-5rqpr"] Oct 13 09:02:18 crc kubenswrapper[4685]: I1013 09:02:18.179828 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5c8fd5c86b-qkv7x"] Oct 13 09:02:18 crc kubenswrapper[4685]: I1013 09:02:18.302330 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"babeefc6-9cbe-4f6c-9487-731a9270c1c4","Type":"ContainerStarted","Data":"dcc4502a4608563d57e3b8098244b876ac27b610f2a0f6a09abd8229da1f5935"} Oct 13 09:02:18 crc kubenswrapper[4685]: I1013 09:02:18.331211 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" event={"ID":"8c48e301-7038-4b0c-9bbf-dda3aa596fb4","Type":"ContainerStarted","Data":"6ba8c4b8aeb6308a79d98c48c79b8f0150e8d2703f0074cadb076d4c70632782"} Oct 13 09:02:18 crc kubenswrapper[4685]: I1013 09:02:18.337963 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e81aae5c-c9f2-455c-8154-fa64d948bcab","Type":"ContainerStarted","Data":"bf89c9a2a3f92e59f2152f7c8cad3c8ee995e7f42b49c371fa94b397af442c4f"} Oct 13 09:02:18 crc kubenswrapper[4685]: I1013 09:02:18.353829 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c8fd5c86b-qkv7x" event={"ID":"89e018d5-d463-4b0e-8fef-f1d3230d703c","Type":"ContainerStarted","Data":"b929f4caf28b25b65a2d81186327ff1794a93ad7e305682a616a94112991ae7b"} Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.380898 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-786f745fcf-5snx2"] Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.385119 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.388177 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.388379 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.393630 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-786f745fcf-5snx2"] Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.462442 4685 generic.go:334] "Generic (PLEG): container finished" podID="8c48e301-7038-4b0c-9bbf-dda3aa596fb4" containerID="8b77488bcc5198c6772e2229565c7cf15c273c176eef7822ac4d22aabb7c87fd" exitCode=0 Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.462505 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" event={"ID":"8c48e301-7038-4b0c-9bbf-dda3aa596fb4","Type":"ContainerDied","Data":"8b77488bcc5198c6772e2229565c7cf15c273c176eef7822ac4d22aabb7c87fd"} Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.488338 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e81aae5c-c9f2-455c-8154-fa64d948bcab","Type":"ContainerStarted","Data":"5410b128f2e6ae710b0ee1323e2c57e92c1d6fee98ce55182f7f60d4a792dee0"} Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.501072 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a0c8086-2865-4107-b424-b7820834b297-internal-tls-certs\") pod \"neutron-786f745fcf-5snx2\" (UID: \"4a0c8086-2865-4107-b424-b7820834b297\") " pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.501102 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4a0c8086-2865-4107-b424-b7820834b297-config\") pod \"neutron-786f745fcf-5snx2\" (UID: \"4a0c8086-2865-4107-b424-b7820834b297\") " pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.501163 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4a0c8086-2865-4107-b424-b7820834b297-httpd-config\") pod \"neutron-786f745fcf-5snx2\" (UID: \"4a0c8086-2865-4107-b424-b7820834b297\") " pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.501202 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxl4z\" (UniqueName: \"kubernetes.io/projected/4a0c8086-2865-4107-b424-b7820834b297-kube-api-access-xxl4z\") pod \"neutron-786f745fcf-5snx2\" (UID: \"4a0c8086-2865-4107-b424-b7820834b297\") " pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.501223 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a0c8086-2865-4107-b424-b7820834b297-public-tls-certs\") pod \"neutron-786f745fcf-5snx2\" (UID: \"4a0c8086-2865-4107-b424-b7820834b297\") " pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.501237 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a0c8086-2865-4107-b424-b7820834b297-ovndb-tls-certs\") pod \"neutron-786f745fcf-5snx2\" (UID: \"4a0c8086-2865-4107-b424-b7820834b297\") " pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.501273 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a0c8086-2865-4107-b424-b7820834b297-combined-ca-bundle\") pod \"neutron-786f745fcf-5snx2\" (UID: \"4a0c8086-2865-4107-b424-b7820834b297\") " pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.565258 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c8fd5c86b-qkv7x" event={"ID":"89e018d5-d463-4b0e-8fef-f1d3230d703c","Type":"ContainerStarted","Data":"aec73af69652acd016a65ef6c7da86c1afc6374c84a7fc156440054ff0191bc3"} Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.565300 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c8fd5c86b-qkv7x" event={"ID":"89e018d5-d463-4b0e-8fef-f1d3230d703c","Type":"ContainerStarted","Data":"39bc7e2bca44e0fe260292a0d1b6b6755d2e47c4bb3345e20a1f5eecbc4e2e70"} Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.566405 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5c8fd5c86b-qkv7x" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.569656 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"babeefc6-9cbe-4f6c-9487-731a9270c1c4","Type":"ContainerStarted","Data":"1151b5de139a4c03b47ac9c2ee15276efe6b1481c2f0cba8e9123177807e5804"} Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.599145 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5c8fd5c86b-qkv7x" podStartSLOduration=3.599131771 podStartE2EDuration="3.599131771s" podCreationTimestamp="2025-10-13 09:02:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:02:19.598087273 +0000 UTC m=+1064.745963034" watchObservedRunningTime="2025-10-13 09:02:19.599131771 +0000 UTC m=+1064.747007532" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.605138 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxl4z\" (UniqueName: \"kubernetes.io/projected/4a0c8086-2865-4107-b424-b7820834b297-kube-api-access-xxl4z\") pod \"neutron-786f745fcf-5snx2\" (UID: \"4a0c8086-2865-4107-b424-b7820834b297\") " pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.605197 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a0c8086-2865-4107-b424-b7820834b297-public-tls-certs\") pod \"neutron-786f745fcf-5snx2\" (UID: \"4a0c8086-2865-4107-b424-b7820834b297\") " pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.605217 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a0c8086-2865-4107-b424-b7820834b297-ovndb-tls-certs\") pod \"neutron-786f745fcf-5snx2\" (UID: \"4a0c8086-2865-4107-b424-b7820834b297\") " pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.605256 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a0c8086-2865-4107-b424-b7820834b297-combined-ca-bundle\") pod \"neutron-786f745fcf-5snx2\" (UID: \"4a0c8086-2865-4107-b424-b7820834b297\") " pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.605348 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a0c8086-2865-4107-b424-b7820834b297-internal-tls-certs\") pod \"neutron-786f745fcf-5snx2\" (UID: \"4a0c8086-2865-4107-b424-b7820834b297\") " pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.605401 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4a0c8086-2865-4107-b424-b7820834b297-config\") pod \"neutron-786f745fcf-5snx2\" (UID: \"4a0c8086-2865-4107-b424-b7820834b297\") " pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.605464 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4a0c8086-2865-4107-b424-b7820834b297-httpd-config\") pod \"neutron-786f745fcf-5snx2\" (UID: \"4a0c8086-2865-4107-b424-b7820834b297\") " pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.616055 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4a0c8086-2865-4107-b424-b7820834b297-httpd-config\") pod \"neutron-786f745fcf-5snx2\" (UID: \"4a0c8086-2865-4107-b424-b7820834b297\") " pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.619580 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a0c8086-2865-4107-b424-b7820834b297-internal-tls-certs\") pod \"neutron-786f745fcf-5snx2\" (UID: \"4a0c8086-2865-4107-b424-b7820834b297\") " pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.627463 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a0c8086-2865-4107-b424-b7820834b297-public-tls-certs\") pod \"neutron-786f745fcf-5snx2\" (UID: \"4a0c8086-2865-4107-b424-b7820834b297\") " pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.629136 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxl4z\" (UniqueName: \"kubernetes.io/projected/4a0c8086-2865-4107-b424-b7820834b297-kube-api-access-xxl4z\") pod \"neutron-786f745fcf-5snx2\" (UID: \"4a0c8086-2865-4107-b424-b7820834b297\") " pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.631653 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a0c8086-2865-4107-b424-b7820834b297-combined-ca-bundle\") pod \"neutron-786f745fcf-5snx2\" (UID: \"4a0c8086-2865-4107-b424-b7820834b297\") " pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.632420 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a0c8086-2865-4107-b424-b7820834b297-ovndb-tls-certs\") pod \"neutron-786f745fcf-5snx2\" (UID: \"4a0c8086-2865-4107-b424-b7820834b297\") " pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.635645 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4a0c8086-2865-4107-b424-b7820834b297-config\") pod \"neutron-786f745fcf-5snx2\" (UID: \"4a0c8086-2865-4107-b424-b7820834b297\") " pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:19 crc kubenswrapper[4685]: I1013 09:02:19.731003 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:20 crc kubenswrapper[4685]: I1013 09:02:20.428566 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-786f745fcf-5snx2"] Oct 13 09:02:20 crc kubenswrapper[4685]: I1013 09:02:20.578976 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-786f745fcf-5snx2" event={"ID":"4a0c8086-2865-4107-b424-b7820834b297","Type":"ContainerStarted","Data":"9b62c1a18983a8bb1c284978a9fa9a4c355472a2d198893e35e7f26371bd9c50"} Oct 13 09:02:20 crc kubenswrapper[4685]: I1013 09:02:20.581445 4685 generic.go:334] "Generic (PLEG): container finished" podID="05506d41-5449-4eb6-86ce-0ac5f5c880ba" containerID="114f444907b28763ac128396f5487639845624f6c74051d43cbdffb59c484704" exitCode=0 Oct 13 09:02:20 crc kubenswrapper[4685]: I1013 09:02:20.581507 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5p6fr" event={"ID":"05506d41-5449-4eb6-86ce-0ac5f5c880ba","Type":"ContainerDied","Data":"114f444907b28763ac128396f5487639845624f6c74051d43cbdffb59c484704"} Oct 13 09:02:20 crc kubenswrapper[4685]: I1013 09:02:20.593531 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" event={"ID":"8c48e301-7038-4b0c-9bbf-dda3aa596fb4","Type":"ContainerStarted","Data":"ea2357dd703bedd260550e5b0fff51e59da36f622be0707c9fc5ff0ff1e87566"} Oct 13 09:02:20 crc kubenswrapper[4685]: I1013 09:02:20.593619 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:20 crc kubenswrapper[4685]: I1013 09:02:20.602269 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e81aae5c-c9f2-455c-8154-fa64d948bcab","Type":"ContainerStarted","Data":"aba2a67b03bcffa3d89c46a94c08cfd4f235737c2fabffcb870c7a5181e0fbde"} Oct 13 09:02:20 crc kubenswrapper[4685]: I1013 09:02:20.606609 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"babeefc6-9cbe-4f6c-9487-731a9270c1c4","Type":"ContainerStarted","Data":"3346630c966f3d7aff2a38529b3649c2ce0e925ee871b7630194486f5d391a51"} Oct 13 09:02:20 crc kubenswrapper[4685]: I1013 09:02:20.620088 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" podStartSLOduration=4.620071589 podStartE2EDuration="4.620071589s" podCreationTimestamp="2025-10-13 09:02:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:02:20.619277238 +0000 UTC m=+1065.767152999" watchObservedRunningTime="2025-10-13 09:02:20.620071589 +0000 UTC m=+1065.767947350" Oct 13 09:02:20 crc kubenswrapper[4685]: I1013 09:02:20.645220 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.645200015 podStartE2EDuration="4.645200015s" podCreationTimestamp="2025-10-13 09:02:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:02:20.63842666 +0000 UTC m=+1065.786302421" watchObservedRunningTime="2025-10-13 09:02:20.645200015 +0000 UTC m=+1065.793075776" Oct 13 09:02:20 crc kubenswrapper[4685]: I1013 09:02:20.668380 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.668357637 podStartE2EDuration="4.668357637s" podCreationTimestamp="2025-10-13 09:02:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:02:20.659072114 +0000 UTC m=+1065.806947875" watchObservedRunningTime="2025-10-13 09:02:20.668357637 +0000 UTC m=+1065.816233398" Oct 13 09:02:21 crc kubenswrapper[4685]: I1013 09:02:21.619147 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-786f745fcf-5snx2" event={"ID":"4a0c8086-2865-4107-b424-b7820834b297","Type":"ContainerStarted","Data":"31f5d19148e532cf3984c02a022087b0ec229ad487a320e6e3a63964308348d1"} Oct 13 09:02:21 crc kubenswrapper[4685]: I1013 09:02:21.619395 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-786f745fcf-5snx2" event={"ID":"4a0c8086-2865-4107-b424-b7820834b297","Type":"ContainerStarted","Data":"b5f4109edc80089f48904e6fd39cf58b8b1b580ad7e68065109ce2ac2f932ead"} Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.058819 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5p6fr" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.078228 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-786f745fcf-5snx2" podStartSLOduration=3.07820809 podStartE2EDuration="3.07820809s" podCreationTimestamp="2025-10-13 09:02:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:02:21.637391008 +0000 UTC m=+1066.785266769" watchObservedRunningTime="2025-10-13 09:02:22.07820809 +0000 UTC m=+1067.226083851" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.165948 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/05506d41-5449-4eb6-86ce-0ac5f5c880ba-db-sync-config-data\") pod \"05506d41-5449-4eb6-86ce-0ac5f5c880ba\" (UID: \"05506d41-5449-4eb6-86ce-0ac5f5c880ba\") " Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.166112 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l44f9\" (UniqueName: \"kubernetes.io/projected/05506d41-5449-4eb6-86ce-0ac5f5c880ba-kube-api-access-l44f9\") pod \"05506d41-5449-4eb6-86ce-0ac5f5c880ba\" (UID: \"05506d41-5449-4eb6-86ce-0ac5f5c880ba\") " Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.166145 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05506d41-5449-4eb6-86ce-0ac5f5c880ba-combined-ca-bundle\") pod \"05506d41-5449-4eb6-86ce-0ac5f5c880ba\" (UID: \"05506d41-5449-4eb6-86ce-0ac5f5c880ba\") " Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.171476 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05506d41-5449-4eb6-86ce-0ac5f5c880ba-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "05506d41-5449-4eb6-86ce-0ac5f5c880ba" (UID: "05506d41-5449-4eb6-86ce-0ac5f5c880ba"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.172183 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05506d41-5449-4eb6-86ce-0ac5f5c880ba-kube-api-access-l44f9" (OuterVolumeSpecName: "kube-api-access-l44f9") pod "05506d41-5449-4eb6-86ce-0ac5f5c880ba" (UID: "05506d41-5449-4eb6-86ce-0ac5f5c880ba"). InnerVolumeSpecName "kube-api-access-l44f9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.204625 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05506d41-5449-4eb6-86ce-0ac5f5c880ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "05506d41-5449-4eb6-86ce-0ac5f5c880ba" (UID: "05506d41-5449-4eb6-86ce-0ac5f5c880ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.268521 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05506d41-5449-4eb6-86ce-0ac5f5c880ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.268569 4685 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/05506d41-5449-4eb6-86ce-0ac5f5c880ba-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.268582 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l44f9\" (UniqueName: \"kubernetes.io/projected/05506d41-5449-4eb6-86ce-0ac5f5c880ba-kube-api-access-l44f9\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.647029 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5p6fr" event={"ID":"05506d41-5449-4eb6-86ce-0ac5f5c880ba","Type":"ContainerDied","Data":"8e9e8ef00358a7eabc8d68ddaf92adcb440ded6caecc8df0037c974b88e20885"} Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.647064 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5p6fr" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.647087 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e9e8ef00358a7eabc8d68ddaf92adcb440ded6caecc8df0037c974b88e20885" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.648349 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.853749 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-58b69dbf78-cznmg"] Oct 13 09:02:22 crc kubenswrapper[4685]: E1013 09:02:22.854196 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05506d41-5449-4eb6-86ce-0ac5f5c880ba" containerName="barbican-db-sync" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.854215 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="05506d41-5449-4eb6-86ce-0ac5f5c880ba" containerName="barbican-db-sync" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.854397 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="05506d41-5449-4eb6-86ce-0ac5f5c880ba" containerName="barbican-db-sync" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.863780 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-58b69dbf78-cznmg" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.870739 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-b4jjn" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.875954 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.881431 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.925428 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-58b69dbf78-cznmg"] Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.955409 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-f7b8c695-sw5hg"] Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.957016 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-f7b8c695-sw5hg" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.960453 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.966400 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-f7b8c695-sw5hg"] Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.981714 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad40cbfc-60f4-4ff0-9106-90a9941d5c10-logs\") pod \"barbican-keystone-listener-58b69dbf78-cznmg\" (UID: \"ad40cbfc-60f4-4ff0-9106-90a9941d5c10\") " pod="openstack/barbican-keystone-listener-58b69dbf78-cznmg" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.981786 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spfdr\" (UniqueName: \"kubernetes.io/projected/ad40cbfc-60f4-4ff0-9106-90a9941d5c10-kube-api-access-spfdr\") pod \"barbican-keystone-listener-58b69dbf78-cznmg\" (UID: \"ad40cbfc-60f4-4ff0-9106-90a9941d5c10\") " pod="openstack/barbican-keystone-listener-58b69dbf78-cznmg" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.981821 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad40cbfc-60f4-4ff0-9106-90a9941d5c10-config-data-custom\") pod \"barbican-keystone-listener-58b69dbf78-cznmg\" (UID: \"ad40cbfc-60f4-4ff0-9106-90a9941d5c10\") " pod="openstack/barbican-keystone-listener-58b69dbf78-cznmg" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.981894 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad40cbfc-60f4-4ff0-9106-90a9941d5c10-config-data\") pod \"barbican-keystone-listener-58b69dbf78-cznmg\" (UID: \"ad40cbfc-60f4-4ff0-9106-90a9941d5c10\") " pod="openstack/barbican-keystone-listener-58b69dbf78-cznmg" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.981925 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad40cbfc-60f4-4ff0-9106-90a9941d5c10-combined-ca-bundle\") pod \"barbican-keystone-listener-58b69dbf78-cznmg\" (UID: \"ad40cbfc-60f4-4ff0-9106-90a9941d5c10\") " pod="openstack/barbican-keystone-listener-58b69dbf78-cznmg" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.983073 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.983139 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.983206 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.983998 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"722c93d4dbc8e9002b13fe5143e24de1b4add33c78bcd87a264b498be58dd6fd"} pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 09:02:22 crc kubenswrapper[4685]: I1013 09:02:22.984062 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" containerID="cri-o://722c93d4dbc8e9002b13fe5143e24de1b4add33c78bcd87a264b498be58dd6fd" gracePeriod=600 Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.041744 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-5rqpr"] Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.042012 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" podUID="8c48e301-7038-4b0c-9bbf-dda3aa596fb4" containerName="dnsmasq-dns" containerID="cri-o://ea2357dd703bedd260550e5b0fff51e59da36f622be0707c9fc5ff0ff1e87566" gracePeriod=10 Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.049551 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.049854 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.084980 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11-combined-ca-bundle\") pod \"barbican-worker-f7b8c695-sw5hg\" (UID: \"102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11\") " pod="openstack/barbican-worker-f7b8c695-sw5hg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.085055 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad40cbfc-60f4-4ff0-9106-90a9941d5c10-logs\") pod \"barbican-keystone-listener-58b69dbf78-cznmg\" (UID: \"ad40cbfc-60f4-4ff0-9106-90a9941d5c10\") " pod="openstack/barbican-keystone-listener-58b69dbf78-cznmg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.085155 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spfdr\" (UniqueName: \"kubernetes.io/projected/ad40cbfc-60f4-4ff0-9106-90a9941d5c10-kube-api-access-spfdr\") pod \"barbican-keystone-listener-58b69dbf78-cznmg\" (UID: \"ad40cbfc-60f4-4ff0-9106-90a9941d5c10\") " pod="openstack/barbican-keystone-listener-58b69dbf78-cznmg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.085220 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad40cbfc-60f4-4ff0-9106-90a9941d5c10-config-data-custom\") pod \"barbican-keystone-listener-58b69dbf78-cznmg\" (UID: \"ad40cbfc-60f4-4ff0-9106-90a9941d5c10\") " pod="openstack/barbican-keystone-listener-58b69dbf78-cznmg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.085258 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11-config-data-custom\") pod \"barbican-worker-f7b8c695-sw5hg\" (UID: \"102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11\") " pod="openstack/barbican-worker-f7b8c695-sw5hg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.085340 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgrhq\" (UniqueName: \"kubernetes.io/projected/102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11-kube-api-access-cgrhq\") pod \"barbican-worker-f7b8c695-sw5hg\" (UID: \"102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11\") " pod="openstack/barbican-worker-f7b8c695-sw5hg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.085378 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11-config-data\") pod \"barbican-worker-f7b8c695-sw5hg\" (UID: \"102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11\") " pod="openstack/barbican-worker-f7b8c695-sw5hg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.085456 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad40cbfc-60f4-4ff0-9106-90a9941d5c10-config-data\") pod \"barbican-keystone-listener-58b69dbf78-cznmg\" (UID: \"ad40cbfc-60f4-4ff0-9106-90a9941d5c10\") " pod="openstack/barbican-keystone-listener-58b69dbf78-cznmg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.085497 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad40cbfc-60f4-4ff0-9106-90a9941d5c10-combined-ca-bundle\") pod \"barbican-keystone-listener-58b69dbf78-cznmg\" (UID: \"ad40cbfc-60f4-4ff0-9106-90a9941d5c10\") " pod="openstack/barbican-keystone-listener-58b69dbf78-cznmg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.085532 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11-logs\") pod \"barbican-worker-f7b8c695-sw5hg\" (UID: \"102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11\") " pod="openstack/barbican-worker-f7b8c695-sw5hg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.087389 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad40cbfc-60f4-4ff0-9106-90a9941d5c10-logs\") pod \"barbican-keystone-listener-58b69dbf78-cznmg\" (UID: \"ad40cbfc-60f4-4ff0-9106-90a9941d5c10\") " pod="openstack/barbican-keystone-listener-58b69dbf78-cznmg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.094770 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad40cbfc-60f4-4ff0-9106-90a9941d5c10-combined-ca-bundle\") pod \"barbican-keystone-listener-58b69dbf78-cznmg\" (UID: \"ad40cbfc-60f4-4ff0-9106-90a9941d5c10\") " pod="openstack/barbican-keystone-listener-58b69dbf78-cznmg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.097878 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-n8774"] Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.106374 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.098075 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad40cbfc-60f4-4ff0-9106-90a9941d5c10-config-data\") pod \"barbican-keystone-listener-58b69dbf78-cznmg\" (UID: \"ad40cbfc-60f4-4ff0-9106-90a9941d5c10\") " pod="openstack/barbican-keystone-listener-58b69dbf78-cznmg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.102967 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad40cbfc-60f4-4ff0-9106-90a9941d5c10-config-data-custom\") pod \"barbican-keystone-listener-58b69dbf78-cznmg\" (UID: \"ad40cbfc-60f4-4ff0-9106-90a9941d5c10\") " pod="openstack/barbican-keystone-listener-58b69dbf78-cznmg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.163318 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spfdr\" (UniqueName: \"kubernetes.io/projected/ad40cbfc-60f4-4ff0-9106-90a9941d5c10-kube-api-access-spfdr\") pod \"barbican-keystone-listener-58b69dbf78-cznmg\" (UID: \"ad40cbfc-60f4-4ff0-9106-90a9941d5c10\") " pod="openstack/barbican-keystone-listener-58b69dbf78-cznmg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.180996 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-n8774"] Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.183565 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-58b69dbf78-cznmg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.186880 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-n8774\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.187069 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11-config-data-custom\") pod \"barbican-worker-f7b8c695-sw5hg\" (UID: \"102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11\") " pod="openstack/barbican-worker-f7b8c695-sw5hg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.187307 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-n8774\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.187393 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssdrm\" (UniqueName: \"kubernetes.io/projected/3688036a-1c46-45c2-ac9b-f2451a738f17-kube-api-access-ssdrm\") pod \"dnsmasq-dns-848cf88cfc-n8774\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.187507 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-n8774\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.187613 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgrhq\" (UniqueName: \"kubernetes.io/projected/102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11-kube-api-access-cgrhq\") pod \"barbican-worker-f7b8c695-sw5hg\" (UID: \"102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11\") " pod="openstack/barbican-worker-f7b8c695-sw5hg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.187692 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-n8774\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.187777 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11-config-data\") pod \"barbican-worker-f7b8c695-sw5hg\" (UID: \"102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11\") " pod="openstack/barbican-worker-f7b8c695-sw5hg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.187870 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11-logs\") pod \"barbican-worker-f7b8c695-sw5hg\" (UID: \"102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11\") " pod="openstack/barbican-worker-f7b8c695-sw5hg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.187980 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11-combined-ca-bundle\") pod \"barbican-worker-f7b8c695-sw5hg\" (UID: \"102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11\") " pod="openstack/barbican-worker-f7b8c695-sw5hg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.188139 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-config\") pod \"dnsmasq-dns-848cf88cfc-n8774\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.192335 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11-logs\") pod \"barbican-worker-f7b8c695-sw5hg\" (UID: \"102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11\") " pod="openstack/barbican-worker-f7b8c695-sw5hg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.193640 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11-combined-ca-bundle\") pod \"barbican-worker-f7b8c695-sw5hg\" (UID: \"102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11\") " pod="openstack/barbican-worker-f7b8c695-sw5hg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.201129 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11-config-data-custom\") pod \"barbican-worker-f7b8c695-sw5hg\" (UID: \"102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11\") " pod="openstack/barbican-worker-f7b8c695-sw5hg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.225594 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgrhq\" (UniqueName: \"kubernetes.io/projected/102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11-kube-api-access-cgrhq\") pod \"barbican-worker-f7b8c695-sw5hg\" (UID: \"102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11\") " pod="openstack/barbican-worker-f7b8c695-sw5hg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.226168 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11-config-data\") pod \"barbican-worker-f7b8c695-sw5hg\" (UID: \"102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11\") " pod="openstack/barbican-worker-f7b8c695-sw5hg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.280060 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-f7b8c695-sw5hg" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.289585 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-config\") pod \"dnsmasq-dns-848cf88cfc-n8774\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.289637 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-n8774\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.289681 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-n8774\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.289705 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssdrm\" (UniqueName: \"kubernetes.io/projected/3688036a-1c46-45c2-ac9b-f2451a738f17-kube-api-access-ssdrm\") pod \"dnsmasq-dns-848cf88cfc-n8774\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.289737 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-n8774\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.289762 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-n8774\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.290692 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-config\") pod \"dnsmasq-dns-848cf88cfc-n8774\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.290807 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-n8774\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.290846 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-n8774\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.291341 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-n8774\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.292461 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-n8774\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.326828 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssdrm\" (UniqueName: \"kubernetes.io/projected/3688036a-1c46-45c2-ac9b-f2451a738f17-kube-api-access-ssdrm\") pod \"dnsmasq-dns-848cf88cfc-n8774\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.361731 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5746d5f994-pjpl9"] Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.363127 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.372336 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.392568 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9dd392f6-cf1c-4d69-81f8-167718c41926-config-data-custom\") pod \"barbican-api-5746d5f994-pjpl9\" (UID: \"9dd392f6-cf1c-4d69-81f8-167718c41926\") " pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.392621 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dd392f6-cf1c-4d69-81f8-167718c41926-combined-ca-bundle\") pod \"barbican-api-5746d5f994-pjpl9\" (UID: \"9dd392f6-cf1c-4d69-81f8-167718c41926\") " pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.392715 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dd392f6-cf1c-4d69-81f8-167718c41926-config-data\") pod \"barbican-api-5746d5f994-pjpl9\" (UID: \"9dd392f6-cf1c-4d69-81f8-167718c41926\") " pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.392771 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dd392f6-cf1c-4d69-81f8-167718c41926-logs\") pod \"barbican-api-5746d5f994-pjpl9\" (UID: \"9dd392f6-cf1c-4d69-81f8-167718c41926\") " pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.392790 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nkdq\" (UniqueName: \"kubernetes.io/projected/9dd392f6-cf1c-4d69-81f8-167718c41926-kube-api-access-8nkdq\") pod \"barbican-api-5746d5f994-pjpl9\" (UID: \"9dd392f6-cf1c-4d69-81f8-167718c41926\") " pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.406250 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5746d5f994-pjpl9"] Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.426656 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.494479 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dd392f6-cf1c-4d69-81f8-167718c41926-combined-ca-bundle\") pod \"barbican-api-5746d5f994-pjpl9\" (UID: \"9dd392f6-cf1c-4d69-81f8-167718c41926\") " pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.495382 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dd392f6-cf1c-4d69-81f8-167718c41926-config-data\") pod \"barbican-api-5746d5f994-pjpl9\" (UID: \"9dd392f6-cf1c-4d69-81f8-167718c41926\") " pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.495505 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dd392f6-cf1c-4d69-81f8-167718c41926-logs\") pod \"barbican-api-5746d5f994-pjpl9\" (UID: \"9dd392f6-cf1c-4d69-81f8-167718c41926\") " pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.495529 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nkdq\" (UniqueName: \"kubernetes.io/projected/9dd392f6-cf1c-4d69-81f8-167718c41926-kube-api-access-8nkdq\") pod \"barbican-api-5746d5f994-pjpl9\" (UID: \"9dd392f6-cf1c-4d69-81f8-167718c41926\") " pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.495573 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9dd392f6-cf1c-4d69-81f8-167718c41926-config-data-custom\") pod \"barbican-api-5746d5f994-pjpl9\" (UID: \"9dd392f6-cf1c-4d69-81f8-167718c41926\") " pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.497476 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dd392f6-cf1c-4d69-81f8-167718c41926-logs\") pod \"barbican-api-5746d5f994-pjpl9\" (UID: \"9dd392f6-cf1c-4d69-81f8-167718c41926\") " pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.509044 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dd392f6-cf1c-4d69-81f8-167718c41926-combined-ca-bundle\") pod \"barbican-api-5746d5f994-pjpl9\" (UID: \"9dd392f6-cf1c-4d69-81f8-167718c41926\") " pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.517853 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9dd392f6-cf1c-4d69-81f8-167718c41926-config-data-custom\") pod \"barbican-api-5746d5f994-pjpl9\" (UID: \"9dd392f6-cf1c-4d69-81f8-167718c41926\") " pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.526461 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dd392f6-cf1c-4d69-81f8-167718c41926-config-data\") pod \"barbican-api-5746d5f994-pjpl9\" (UID: \"9dd392f6-cf1c-4d69-81f8-167718c41926\") " pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.554243 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nkdq\" (UniqueName: \"kubernetes.io/projected/9dd392f6-cf1c-4d69-81f8-167718c41926-kube-api-access-8nkdq\") pod \"barbican-api-5746d5f994-pjpl9\" (UID: \"9dd392f6-cf1c-4d69-81f8-167718c41926\") " pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.661450 4685 generic.go:334] "Generic (PLEG): container finished" podID="8c48e301-7038-4b0c-9bbf-dda3aa596fb4" containerID="ea2357dd703bedd260550e5b0fff51e59da36f622be0707c9fc5ff0ff1e87566" exitCode=0 Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.661507 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" event={"ID":"8c48e301-7038-4b0c-9bbf-dda3aa596fb4","Type":"ContainerDied","Data":"ea2357dd703bedd260550e5b0fff51e59da36f622be0707c9fc5ff0ff1e87566"} Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.662518 4685 generic.go:334] "Generic (PLEG): container finished" podID="40fbc731-ab15-4cc0-9eab-1ec6c0848cbc" containerID="16c5d0f261753dbe0cc387ffa98a76a6a4a86536f143165e934db71b95f53192" exitCode=0 Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.662569 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b29nc" event={"ID":"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc","Type":"ContainerDied","Data":"16c5d0f261753dbe0cc387ffa98a76a6a4a86536f143165e934db71b95f53192"} Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.682594 4685 generic.go:334] "Generic (PLEG): container finished" podID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerID="722c93d4dbc8e9002b13fe5143e24de1b4add33c78bcd87a264b498be58dd6fd" exitCode=0 Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.683446 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerDied","Data":"722c93d4dbc8e9002b13fe5143e24de1b4add33c78bcd87a264b498be58dd6fd"} Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.683481 4685 scope.go:117] "RemoveContainer" containerID="c54513235e556be91a1895a9b0943234dc7b6e87cbf946c33f9dc14ea2f61818" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.755098 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:23 crc kubenswrapper[4685]: I1013 09:02:23.792623 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-db976b79d-q446d" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.784267 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7cc9c86456-5hsl7"] Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.788258 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.794353 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.794616 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.805438 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7cc9c86456-5hsl7"] Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.842718 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27d72d0b-473c-4983-be92-7b3eca7d5c45-logs\") pod \"barbican-api-7cc9c86456-5hsl7\" (UID: \"27d72d0b-473c-4983-be92-7b3eca7d5c45\") " pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.842838 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27d72d0b-473c-4983-be92-7b3eca7d5c45-public-tls-certs\") pod \"barbican-api-7cc9c86456-5hsl7\" (UID: \"27d72d0b-473c-4983-be92-7b3eca7d5c45\") " pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.842879 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spdsr\" (UniqueName: \"kubernetes.io/projected/27d72d0b-473c-4983-be92-7b3eca7d5c45-kube-api-access-spdsr\") pod \"barbican-api-7cc9c86456-5hsl7\" (UID: \"27d72d0b-473c-4983-be92-7b3eca7d5c45\") " pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.842938 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d72d0b-473c-4983-be92-7b3eca7d5c45-combined-ca-bundle\") pod \"barbican-api-7cc9c86456-5hsl7\" (UID: \"27d72d0b-473c-4983-be92-7b3eca7d5c45\") " pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.842962 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27d72d0b-473c-4983-be92-7b3eca7d5c45-config-data\") pod \"barbican-api-7cc9c86456-5hsl7\" (UID: \"27d72d0b-473c-4983-be92-7b3eca7d5c45\") " pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.843010 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27d72d0b-473c-4983-be92-7b3eca7d5c45-internal-tls-certs\") pod \"barbican-api-7cc9c86456-5hsl7\" (UID: \"27d72d0b-473c-4983-be92-7b3eca7d5c45\") " pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.843034 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/27d72d0b-473c-4983-be92-7b3eca7d5c45-config-data-custom\") pod \"barbican-api-7cc9c86456-5hsl7\" (UID: \"27d72d0b-473c-4983-be92-7b3eca7d5c45\") " pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.945231 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d72d0b-473c-4983-be92-7b3eca7d5c45-combined-ca-bundle\") pod \"barbican-api-7cc9c86456-5hsl7\" (UID: \"27d72d0b-473c-4983-be92-7b3eca7d5c45\") " pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.945295 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27d72d0b-473c-4983-be92-7b3eca7d5c45-config-data\") pod \"barbican-api-7cc9c86456-5hsl7\" (UID: \"27d72d0b-473c-4983-be92-7b3eca7d5c45\") " pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.945331 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27d72d0b-473c-4983-be92-7b3eca7d5c45-internal-tls-certs\") pod \"barbican-api-7cc9c86456-5hsl7\" (UID: \"27d72d0b-473c-4983-be92-7b3eca7d5c45\") " pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.945358 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/27d72d0b-473c-4983-be92-7b3eca7d5c45-config-data-custom\") pod \"barbican-api-7cc9c86456-5hsl7\" (UID: \"27d72d0b-473c-4983-be92-7b3eca7d5c45\") " pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.945398 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27d72d0b-473c-4983-be92-7b3eca7d5c45-logs\") pod \"barbican-api-7cc9c86456-5hsl7\" (UID: \"27d72d0b-473c-4983-be92-7b3eca7d5c45\") " pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.945468 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27d72d0b-473c-4983-be92-7b3eca7d5c45-public-tls-certs\") pod \"barbican-api-7cc9c86456-5hsl7\" (UID: \"27d72d0b-473c-4983-be92-7b3eca7d5c45\") " pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.945492 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spdsr\" (UniqueName: \"kubernetes.io/projected/27d72d0b-473c-4983-be92-7b3eca7d5c45-kube-api-access-spdsr\") pod \"barbican-api-7cc9c86456-5hsl7\" (UID: \"27d72d0b-473c-4983-be92-7b3eca7d5c45\") " pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.946286 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27d72d0b-473c-4983-be92-7b3eca7d5c45-logs\") pod \"barbican-api-7cc9c86456-5hsl7\" (UID: \"27d72d0b-473c-4983-be92-7b3eca7d5c45\") " pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.952314 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27d72d0b-473c-4983-be92-7b3eca7d5c45-internal-tls-certs\") pod \"barbican-api-7cc9c86456-5hsl7\" (UID: \"27d72d0b-473c-4983-be92-7b3eca7d5c45\") " pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.956704 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27d72d0b-473c-4983-be92-7b3eca7d5c45-config-data\") pod \"barbican-api-7cc9c86456-5hsl7\" (UID: \"27d72d0b-473c-4983-be92-7b3eca7d5c45\") " pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.960861 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spdsr\" (UniqueName: \"kubernetes.io/projected/27d72d0b-473c-4983-be92-7b3eca7d5c45-kube-api-access-spdsr\") pod \"barbican-api-7cc9c86456-5hsl7\" (UID: \"27d72d0b-473c-4983-be92-7b3eca7d5c45\") " pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.972839 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27d72d0b-473c-4983-be92-7b3eca7d5c45-public-tls-certs\") pod \"barbican-api-7cc9c86456-5hsl7\" (UID: \"27d72d0b-473c-4983-be92-7b3eca7d5c45\") " pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.976499 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/27d72d0b-473c-4983-be92-7b3eca7d5c45-config-data-custom\") pod \"barbican-api-7cc9c86456-5hsl7\" (UID: \"27d72d0b-473c-4983-be92-7b3eca7d5c45\") " pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:25 crc kubenswrapper[4685]: I1013 09:02:25.981096 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d72d0b-473c-4983-be92-7b3eca7d5c45-combined-ca-bundle\") pod \"barbican-api-7cc9c86456-5hsl7\" (UID: \"27d72d0b-473c-4983-be92-7b3eca7d5c45\") " pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:26 crc kubenswrapper[4685]: I1013 09:02:26.109608 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:26 crc kubenswrapper[4685]: I1013 09:02:26.414699 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-db976b79d-q446d" Oct 13 09:02:26 crc kubenswrapper[4685]: I1013 09:02:26.780050 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 13 09:02:26 crc kubenswrapper[4685]: I1013 09:02:26.780399 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 13 09:02:26 crc kubenswrapper[4685]: I1013 09:02:26.833598 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 13 09:02:26 crc kubenswrapper[4685]: I1013 09:02:26.834569 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 13 09:02:26 crc kubenswrapper[4685]: I1013 09:02:26.968381 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 13 09:02:26 crc kubenswrapper[4685]: I1013 09:02:26.972202 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 13 09:02:27 crc kubenswrapper[4685]: I1013 09:02:27.008573 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 13 09:02:27 crc kubenswrapper[4685]: I1013 09:02:27.032163 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 13 09:02:27 crc kubenswrapper[4685]: I1013 09:02:27.182747 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" podUID="8c48e301-7038-4b0c-9bbf-dda3aa596fb4" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.157:5353: connect: connection refused" Oct 13 09:02:27 crc kubenswrapper[4685]: I1013 09:02:27.741333 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 13 09:02:27 crc kubenswrapper[4685]: I1013 09:02:27.741379 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 13 09:02:27 crc kubenswrapper[4685]: I1013 09:02:27.741391 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 13 09:02:27 crc kubenswrapper[4685]: I1013 09:02:27.741402 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 13 09:02:28 crc kubenswrapper[4685]: I1013 09:02:28.926250 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b29nc" Oct 13 09:02:29 crc kubenswrapper[4685]: I1013 09:02:29.035470 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-combined-ca-bundle\") pod \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " Oct 13 09:02:29 crc kubenswrapper[4685]: I1013 09:02:29.035544 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-scripts\") pod \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " Oct 13 09:02:29 crc kubenswrapper[4685]: I1013 09:02:29.035622 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-fernet-keys\") pod \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " Oct 13 09:02:29 crc kubenswrapper[4685]: I1013 09:02:29.035724 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tpb4\" (UniqueName: \"kubernetes.io/projected/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-kube-api-access-5tpb4\") pod \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " Oct 13 09:02:29 crc kubenswrapper[4685]: I1013 09:02:29.035816 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-config-data\") pod \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " Oct 13 09:02:29 crc kubenswrapper[4685]: I1013 09:02:29.035874 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-credential-keys\") pod \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\" (UID: \"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc\") " Oct 13 09:02:29 crc kubenswrapper[4685]: I1013 09:02:29.052120 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-kube-api-access-5tpb4" (OuterVolumeSpecName: "kube-api-access-5tpb4") pod "40fbc731-ab15-4cc0-9eab-1ec6c0848cbc" (UID: "40fbc731-ab15-4cc0-9eab-1ec6c0848cbc"). InnerVolumeSpecName "kube-api-access-5tpb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:02:29 crc kubenswrapper[4685]: I1013 09:02:29.052673 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "40fbc731-ab15-4cc0-9eab-1ec6c0848cbc" (UID: "40fbc731-ab15-4cc0-9eab-1ec6c0848cbc"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:29 crc kubenswrapper[4685]: I1013 09:02:29.055559 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-scripts" (OuterVolumeSpecName: "scripts") pod "40fbc731-ab15-4cc0-9eab-1ec6c0848cbc" (UID: "40fbc731-ab15-4cc0-9eab-1ec6c0848cbc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:29 crc kubenswrapper[4685]: I1013 09:02:29.057675 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "40fbc731-ab15-4cc0-9eab-1ec6c0848cbc" (UID: "40fbc731-ab15-4cc0-9eab-1ec6c0848cbc"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:29 crc kubenswrapper[4685]: I1013 09:02:29.121787 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "40fbc731-ab15-4cc0-9eab-1ec6c0848cbc" (UID: "40fbc731-ab15-4cc0-9eab-1ec6c0848cbc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:29 crc kubenswrapper[4685]: I1013 09:02:29.134152 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-config-data" (OuterVolumeSpecName: "config-data") pod "40fbc731-ab15-4cc0-9eab-1ec6c0848cbc" (UID: "40fbc731-ab15-4cc0-9eab-1ec6c0848cbc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:29 crc kubenswrapper[4685]: I1013 09:02:29.137766 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tpb4\" (UniqueName: \"kubernetes.io/projected/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-kube-api-access-5tpb4\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:29 crc kubenswrapper[4685]: I1013 09:02:29.137799 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:29 crc kubenswrapper[4685]: I1013 09:02:29.137809 4685 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:29 crc kubenswrapper[4685]: I1013 09:02:29.137818 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:29 crc kubenswrapper[4685]: I1013 09:02:29.137826 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:29 crc kubenswrapper[4685]: I1013 09:02:29.137836 4685 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:29 crc kubenswrapper[4685]: I1013 09:02:29.765613 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b29nc" event={"ID":"40fbc731-ab15-4cc0-9eab-1ec6c0848cbc","Type":"ContainerDied","Data":"4b103562677017cda82e8fbbc57462022313bf928ac9f7711e1daaba740458ac"} Oct 13 09:02:29 crc kubenswrapper[4685]: I1013 09:02:29.765885 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b103562677017cda82e8fbbc57462022313bf928ac9f7711e1daaba740458ac" Oct 13 09:02:29 crc kubenswrapper[4685]: I1013 09:02:29.765699 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b29nc" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.061701 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-79b7b9958d-7s7c7"] Oct 13 09:02:30 crc kubenswrapper[4685]: E1013 09:02:30.066348 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40fbc731-ab15-4cc0-9eab-1ec6c0848cbc" containerName="keystone-bootstrap" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.066365 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="40fbc731-ab15-4cc0-9eab-1ec6c0848cbc" containerName="keystone-bootstrap" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.066544 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="40fbc731-ab15-4cc0-9eab-1ec6c0848cbc" containerName="keystone-bootstrap" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.067143 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.071581 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.071792 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.071959 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.072107 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.072123 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.072263 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5xp9z" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.097971 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-79b7b9958d-7s7c7"] Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.166059 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/869b9e65-4e24-4046-9b64-c1ad8d970be4-credential-keys\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.166213 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/869b9e65-4e24-4046-9b64-c1ad8d970be4-public-tls-certs\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.166343 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/869b9e65-4e24-4046-9b64-c1ad8d970be4-config-data\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.166488 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/869b9e65-4e24-4046-9b64-c1ad8d970be4-fernet-keys\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.166561 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/869b9e65-4e24-4046-9b64-c1ad8d970be4-internal-tls-certs\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.166648 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnkd2\" (UniqueName: \"kubernetes.io/projected/869b9e65-4e24-4046-9b64-c1ad8d970be4-kube-api-access-jnkd2\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.166678 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/869b9e65-4e24-4046-9b64-c1ad8d970be4-combined-ca-bundle\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.166704 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/869b9e65-4e24-4046-9b64-c1ad8d970be4-scripts\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.268718 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/869b9e65-4e24-4046-9b64-c1ad8d970be4-credential-keys\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.269107 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/869b9e65-4e24-4046-9b64-c1ad8d970be4-public-tls-certs\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.269180 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/869b9e65-4e24-4046-9b64-c1ad8d970be4-config-data\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.269235 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/869b9e65-4e24-4046-9b64-c1ad8d970be4-fernet-keys\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.269291 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/869b9e65-4e24-4046-9b64-c1ad8d970be4-internal-tls-certs\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.269335 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnkd2\" (UniqueName: \"kubernetes.io/projected/869b9e65-4e24-4046-9b64-c1ad8d970be4-kube-api-access-jnkd2\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.269357 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/869b9e65-4e24-4046-9b64-c1ad8d970be4-combined-ca-bundle\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.269378 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/869b9e65-4e24-4046-9b64-c1ad8d970be4-scripts\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.275179 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/869b9e65-4e24-4046-9b64-c1ad8d970be4-internal-tls-certs\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.280630 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/869b9e65-4e24-4046-9b64-c1ad8d970be4-fernet-keys\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.280921 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/869b9e65-4e24-4046-9b64-c1ad8d970be4-scripts\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.286409 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/869b9e65-4e24-4046-9b64-c1ad8d970be4-public-tls-certs\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.288630 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/869b9e65-4e24-4046-9b64-c1ad8d970be4-combined-ca-bundle\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.288652 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/869b9e65-4e24-4046-9b64-c1ad8d970be4-config-data\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.293481 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/869b9e65-4e24-4046-9b64-c1ad8d970be4-credential-keys\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.300472 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnkd2\" (UniqueName: \"kubernetes.io/projected/869b9e65-4e24-4046-9b64-c1ad8d970be4-kube-api-access-jnkd2\") pod \"keystone-79b7b9958d-7s7c7\" (UID: \"869b9e65-4e24-4046-9b64-c1ad8d970be4\") " pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.394475 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.615261 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.794497 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-dns-svc\") pod \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.794775 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-ovsdbserver-nb\") pod \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.794958 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bp67s\" (UniqueName: \"kubernetes.io/projected/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-kube-api-access-bp67s\") pod \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.795071 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-config\") pod \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.795124 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-ovsdbserver-sb\") pod \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.795153 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-dns-swift-storage-0\") pod \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\" (UID: \"8c48e301-7038-4b0c-9bbf-dda3aa596fb4\") " Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.871155 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-kube-api-access-bp67s" (OuterVolumeSpecName: "kube-api-access-bp67s") pod "8c48e301-7038-4b0c-9bbf-dda3aa596fb4" (UID: "8c48e301-7038-4b0c-9bbf-dda3aa596fb4"). InnerVolumeSpecName "kube-api-access-bp67s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.871446 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" event={"ID":"8c48e301-7038-4b0c-9bbf-dda3aa596fb4","Type":"ContainerDied","Data":"6ba8c4b8aeb6308a79d98c48c79b8f0150e8d2703f0074cadb076d4c70632782"} Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.871485 4685 scope.go:117] "RemoveContainer" containerID="ea2357dd703bedd260550e5b0fff51e59da36f622be0707c9fc5ff0ff1e87566" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.871610 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-5rqpr" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.888394 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerStarted","Data":"3648096a0845e8e530c524922c94cf62c1c3b816bc3674fd046d4900ce16a389"} Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.897411 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bp67s\" (UniqueName: \"kubernetes.io/projected/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-kube-api-access-bp67s\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:30 crc kubenswrapper[4685]: I1013 09:02:30.943602 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5746d5f994-pjpl9"] Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.014686 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7cc9c86456-5hsl7"] Oct 13 09:02:31 crc kubenswrapper[4685]: W1013 09:02:31.099822 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27d72d0b_473c_4983_be92_7b3eca7d5c45.slice/crio-f967ea3e88d43d093a474847cef53868a0ffcacb34a9bac157bed356df9f56ac WatchSource:0}: Error finding container f967ea3e88d43d093a474847cef53868a0ffcacb34a9bac157bed356df9f56ac: Status 404 returned error can't find the container with id f967ea3e88d43d093a474847cef53868a0ffcacb34a9bac157bed356df9f56ac Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.278720 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-58b69dbf78-cznmg"] Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.309976 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-f7b8c695-sw5hg"] Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.317557 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-n8774"] Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.375856 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8c48e301-7038-4b0c-9bbf-dda3aa596fb4" (UID: "8c48e301-7038-4b0c-9bbf-dda3aa596fb4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.395404 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8c48e301-7038-4b0c-9bbf-dda3aa596fb4" (UID: "8c48e301-7038-4b0c-9bbf-dda3aa596fb4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.413080 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.413107 4685 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.416605 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-config" (OuterVolumeSpecName: "config") pod "8c48e301-7038-4b0c-9bbf-dda3aa596fb4" (UID: "8c48e301-7038-4b0c-9bbf-dda3aa596fb4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.455719 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8c48e301-7038-4b0c-9bbf-dda3aa596fb4" (UID: "8c48e301-7038-4b0c-9bbf-dda3aa596fb4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.520403 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-config\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.520429 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.573936 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8c48e301-7038-4b0c-9bbf-dda3aa596fb4" (UID: "8c48e301-7038-4b0c-9bbf-dda3aa596fb4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.614330 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-79b7b9958d-7s7c7"] Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.622875 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c48e301-7038-4b0c-9bbf-dda3aa596fb4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.783228 4685 scope.go:117] "RemoveContainer" containerID="8b77488bcc5198c6772e2229565c7cf15c273c176eef7822ac4d22aabb7c87fd" Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.855113 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-5rqpr"] Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.883789 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-5rqpr"] Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.932080 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ffc408c-86d3-4e93-9d49-6ef39cb7d959","Type":"ContainerStarted","Data":"39a2e5629a8f0cc1a200d4b9f131214c0abcf9eb6b5decfd5d264ac00f2ddfcb"} Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.932278 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0ffc408c-86d3-4e93-9d49-6ef39cb7d959" containerName="ceilometer-central-agent" containerID="cri-o://a392b4d4475d36fb966ec8ba4fcd6cb1bd6f904281b206bd0573f0b1765e6cf7" gracePeriod=30 Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.932579 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.932862 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0ffc408c-86d3-4e93-9d49-6ef39cb7d959" containerName="proxy-httpd" containerID="cri-o://39a2e5629a8f0cc1a200d4b9f131214c0abcf9eb6b5decfd5d264ac00f2ddfcb" gracePeriod=30 Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.933020 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0ffc408c-86d3-4e93-9d49-6ef39cb7d959" containerName="sg-core" containerID="cri-o://dfd28dc5a0e2852cf335562850401d5db7a9d724c0154c5edec7099ac902d449" gracePeriod=30 Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.933070 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0ffc408c-86d3-4e93-9d49-6ef39cb7d959" containerName="ceilometer-notification-agent" containerID="cri-o://da0a042d918652a2a33633572db233d9037398e55494df44006a3b65c0286815" gracePeriod=30 Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.941280 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5746d5f994-pjpl9" event={"ID":"9dd392f6-cf1c-4d69-81f8-167718c41926","Type":"ContainerStarted","Data":"cbe6d0a2722732a4d111f053aa7a593a39474ddfd32b4f71164834811c6471d1"} Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.941502 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5746d5f994-pjpl9" event={"ID":"9dd392f6-cf1c-4d69-81f8-167718c41926","Type":"ContainerStarted","Data":"66b2160cade8f8daff442c987cd520da1b43ed981122997765c76f7f7d7a68ef"} Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.943641 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-n8774" event={"ID":"3688036a-1c46-45c2-ac9b-f2451a738f17","Type":"ContainerStarted","Data":"46a6aa94dbe5dd4aa8d17a3e63daccc62c2561e1b1865a4b7a0da44bccd1e07d"} Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.946701 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-f7b8c695-sw5hg" event={"ID":"102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11","Type":"ContainerStarted","Data":"f37b785dc2305e7c66edb5cee44af1fd1086a127ee0a5c5ec6be01311e28b0b6"} Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.955059 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-58b69dbf78-cznmg" event={"ID":"ad40cbfc-60f4-4ff0-9106-90a9941d5c10","Type":"ContainerStarted","Data":"c66273e514a54785cabd8a00fdb822c1df02564cc5b7c554b047c165e9700a1a"} Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.967445 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.140166262 podStartE2EDuration="1m5.967430516s" podCreationTimestamp="2025-10-13 09:01:26 +0000 UTC" firstStartedPulling="2025-10-13 09:01:28.68930929 +0000 UTC m=+1013.837185051" lastFinishedPulling="2025-10-13 09:02:30.516573544 +0000 UTC m=+1075.664449305" observedRunningTime="2025-10-13 09:02:31.955379807 +0000 UTC m=+1077.103255558" watchObservedRunningTime="2025-10-13 09:02:31.967430516 +0000 UTC m=+1077.115306277" Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.975415 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-79b7b9958d-7s7c7" event={"ID":"869b9e65-4e24-4046-9b64-c1ad8d970be4","Type":"ContainerStarted","Data":"8d973317e1c7def357db8468aebef95e43dc41e058ec08e8fe0a324e506cd3fc"} Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.979464 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7cc9c86456-5hsl7" event={"ID":"27d72d0b-473c-4983-be92-7b3eca7d5c45","Type":"ContainerStarted","Data":"5ef511c86d6083a09b23345ec3bebf2997a39c685ee85df81d0c7682e3f26cdb"} Oct 13 09:02:31 crc kubenswrapper[4685]: I1013 09:02:31.979495 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7cc9c86456-5hsl7" event={"ID":"27d72d0b-473c-4983-be92-7b3eca7d5c45","Type":"ContainerStarted","Data":"f967ea3e88d43d093a474847cef53868a0ffcacb34a9bac157bed356df9f56ac"} Oct 13 09:02:32 crc kubenswrapper[4685]: E1013 09:02:32.890896 4685 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod126e6e95_80d8_4e68_a541_8e53a4f5b862.slice/crio-conmon-b115917cd44da7ffaccbe91d120ee61d59a4460f53f619a7e3e0215ca2922259.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod126e6e95_80d8_4e68_a541_8e53a4f5b862.slice/crio-conmon-057d5f72232fe1033e30b206d78832bccc56b3093599e6d33773993e1467b3c1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod126e6e95_80d8_4e68_a541_8e53a4f5b862.slice/crio-057d5f72232fe1033e30b206d78832bccc56b3093599e6d33773993e1467b3c1.scope\": RecentStats: unable to find data in memory cache]" Oct 13 09:02:32 crc kubenswrapper[4685]: I1013 09:02:32.907803 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.039670 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7cc9c86456-5hsl7" event={"ID":"27d72d0b-473c-4983-be92-7b3eca7d5c45","Type":"ContainerStarted","Data":"c024f68f26fa94463f822e3477a44cbfad9f2a10f803717e886a91c8f7646766"} Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.040169 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.040273 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.064261 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-dbcf576b-msgft" podUID="46272601-40bd-43ec-b7a8-6824281972e0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.072033 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7cc9c86456-5hsl7" podStartSLOduration=8.072008437 podStartE2EDuration="8.072008437s" podCreationTimestamp="2025-10-13 09:02:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:02:33.06557402 +0000 UTC m=+1078.213449791" watchObservedRunningTime="2025-10-13 09:02:33.072008437 +0000 UTC m=+1078.219884198" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.077884 4685 generic.go:334] "Generic (PLEG): container finished" podID="0ffc408c-86d3-4e93-9d49-6ef39cb7d959" containerID="39a2e5629a8f0cc1a200d4b9f131214c0abcf9eb6b5decfd5d264ac00f2ddfcb" exitCode=0 Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.078049 4685 generic.go:334] "Generic (PLEG): container finished" podID="0ffc408c-86d3-4e93-9d49-6ef39cb7d959" containerID="dfd28dc5a0e2852cf335562850401d5db7a9d724c0154c5edec7099ac902d449" exitCode=2 Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.078141 4685 generic.go:334] "Generic (PLEG): container finished" podID="0ffc408c-86d3-4e93-9d49-6ef39cb7d959" containerID="a392b4d4475d36fb966ec8ba4fcd6cb1bd6f904281b206bd0573f0b1765e6cf7" exitCode=0 Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.078109 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ffc408c-86d3-4e93-9d49-6ef39cb7d959","Type":"ContainerDied","Data":"39a2e5629a8f0cc1a200d4b9f131214c0abcf9eb6b5decfd5d264ac00f2ddfcb"} Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.078318 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ffc408c-86d3-4e93-9d49-6ef39cb7d959","Type":"ContainerDied","Data":"dfd28dc5a0e2852cf335562850401d5db7a9d724c0154c5edec7099ac902d449"} Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.078412 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ffc408c-86d3-4e93-9d49-6ef39cb7d959","Type":"ContainerDied","Data":"a392b4d4475d36fb966ec8ba4fcd6cb1bd6f904281b206bd0573f0b1765e6cf7"} Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.096420 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5746d5f994-pjpl9" event={"ID":"9dd392f6-cf1c-4d69-81f8-167718c41926","Type":"ContainerStarted","Data":"b559fba90dced4d180fb0896393fea41366c1de9c4acd2e31f5bf95bd3823d3b"} Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.096493 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.096571 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.103791 4685 generic.go:334] "Generic (PLEG): container finished" podID="3688036a-1c46-45c2-ac9b-f2451a738f17" containerID="49805c47194335609a4ffa0973190da27ed6a7b9578213b35d68b415372bef5d" exitCode=0 Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.103854 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-n8774" event={"ID":"3688036a-1c46-45c2-ac9b-f2451a738f17","Type":"ContainerDied","Data":"49805c47194335609a4ffa0973190da27ed6a7b9578213b35d68b415372bef5d"} Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.143871 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5746d5f994-pjpl9" podStartSLOduration=10.143659242 podStartE2EDuration="10.143659242s" podCreationTimestamp="2025-10-13 09:02:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:02:33.130017819 +0000 UTC m=+1078.277893600" watchObservedRunningTime="2025-10-13 09:02:33.143659242 +0000 UTC m=+1078.291535003" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.154573 4685 generic.go:334] "Generic (PLEG): container finished" podID="126e6e95-80d8-4e68-a541-8e53a4f5b862" containerID="b115917cd44da7ffaccbe91d120ee61d59a4460f53f619a7e3e0215ca2922259" exitCode=137 Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.155016 4685 generic.go:334] "Generic (PLEG): container finished" podID="126e6e95-80d8-4e68-a541-8e53a4f5b862" containerID="057d5f72232fe1033e30b206d78832bccc56b3093599e6d33773993e1467b3c1" exitCode=137 Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.155139 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64d69c974d-22pqc" event={"ID":"126e6e95-80d8-4e68-a541-8e53a4f5b862","Type":"ContainerDied","Data":"b115917cd44da7ffaccbe91d120ee61d59a4460f53f619a7e3e0215ca2922259"} Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.155215 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64d69c974d-22pqc" event={"ID":"126e6e95-80d8-4e68-a541-8e53a4f5b862","Type":"ContainerDied","Data":"057d5f72232fe1033e30b206d78832bccc56b3093599e6d33773993e1467b3c1"} Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.212825 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7h6qb" event={"ID":"b225d240-40a0-4d59-aa19-6355515e9f8c","Type":"ContainerStarted","Data":"048241ab2b6e54a12029b77e838f010d8527358e132afb5585f66a15de926667"} Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.218521 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.238689 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-79b7b9958d-7s7c7" event={"ID":"869b9e65-4e24-4046-9b64-c1ad8d970be4","Type":"ContainerStarted","Data":"36178d7cbd839223d704c83c27c1b8d45f7af2b3e3bd0c4f7f5b5b0ec726aa69"} Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.239334 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.269928 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-7h6qb" podStartSLOduration=16.751811783 podStartE2EDuration="58.269891387s" podCreationTimestamp="2025-10-13 09:01:35 +0000 UTC" firstStartedPulling="2025-10-13 09:01:49.064359017 +0000 UTC m=+1034.212234778" lastFinishedPulling="2025-10-13 09:02:30.582438621 +0000 UTC m=+1075.730314382" observedRunningTime="2025-10-13 09:02:33.244256338 +0000 UTC m=+1078.392132109" watchObservedRunningTime="2025-10-13 09:02:33.269891387 +0000 UTC m=+1078.417767148" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.311930 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-79b7b9958d-7s7c7" podStartSLOduration=3.311896424 podStartE2EDuration="3.311896424s" podCreationTimestamp="2025-10-13 09:02:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:02:33.268825539 +0000 UTC m=+1078.416701300" watchObservedRunningTime="2025-10-13 09:02:33.311896424 +0000 UTC m=+1078.459772175" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.367004 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.508175 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvh94\" (UniqueName: \"kubernetes.io/projected/126e6e95-80d8-4e68-a541-8e53a4f5b862-kube-api-access-mvh94\") pod \"126e6e95-80d8-4e68-a541-8e53a4f5b862\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.508506 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/126e6e95-80d8-4e68-a541-8e53a4f5b862-combined-ca-bundle\") pod \"126e6e95-80d8-4e68-a541-8e53a4f5b862\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.508543 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/126e6e95-80d8-4e68-a541-8e53a4f5b862-horizon-tls-certs\") pod \"126e6e95-80d8-4e68-a541-8e53a4f5b862\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.508567 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/126e6e95-80d8-4e68-a541-8e53a4f5b862-horizon-secret-key\") pod \"126e6e95-80d8-4e68-a541-8e53a4f5b862\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.508593 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/126e6e95-80d8-4e68-a541-8e53a4f5b862-scripts\") pod \"126e6e95-80d8-4e68-a541-8e53a4f5b862\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.508630 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/126e6e95-80d8-4e68-a541-8e53a4f5b862-logs\") pod \"126e6e95-80d8-4e68-a541-8e53a4f5b862\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.508691 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/126e6e95-80d8-4e68-a541-8e53a4f5b862-config-data\") pod \"126e6e95-80d8-4e68-a541-8e53a4f5b862\" (UID: \"126e6e95-80d8-4e68-a541-8e53a4f5b862\") " Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.517296 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/126e6e95-80d8-4e68-a541-8e53a4f5b862-kube-api-access-mvh94" (OuterVolumeSpecName: "kube-api-access-mvh94") pod "126e6e95-80d8-4e68-a541-8e53a4f5b862" (UID: "126e6e95-80d8-4e68-a541-8e53a4f5b862"). InnerVolumeSpecName "kube-api-access-mvh94". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.522124 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/126e6e95-80d8-4e68-a541-8e53a4f5b862-logs" (OuterVolumeSpecName: "logs") pod "126e6e95-80d8-4e68-a541-8e53a4f5b862" (UID: "126e6e95-80d8-4e68-a541-8e53a4f5b862"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.527060 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/126e6e95-80d8-4e68-a541-8e53a4f5b862-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "126e6e95-80d8-4e68-a541-8e53a4f5b862" (UID: "126e6e95-80d8-4e68-a541-8e53a4f5b862"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.533176 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c48e301-7038-4b0c-9bbf-dda3aa596fb4" path="/var/lib/kubelet/pods/8c48e301-7038-4b0c-9bbf-dda3aa596fb4/volumes" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.615224 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvh94\" (UniqueName: \"kubernetes.io/projected/126e6e95-80d8-4e68-a541-8e53a4f5b862-kube-api-access-mvh94\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.615495 4685 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/126e6e95-80d8-4e68-a541-8e53a4f5b862-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.615567 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/126e6e95-80d8-4e68-a541-8e53a4f5b862-logs\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.637587 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.637703 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.667282 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/126e6e95-80d8-4e68-a541-8e53a4f5b862-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "126e6e95-80d8-4e68-a541-8e53a4f5b862" (UID: "126e6e95-80d8-4e68-a541-8e53a4f5b862"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.685595 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.685694 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.712328 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/126e6e95-80d8-4e68-a541-8e53a4f5b862-scripts" (OuterVolumeSpecName: "scripts") pod "126e6e95-80d8-4e68-a541-8e53a4f5b862" (UID: "126e6e95-80d8-4e68-a541-8e53a4f5b862"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.717653 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/126e6e95-80d8-4e68-a541-8e53a4f5b862-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.717677 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/126e6e95-80d8-4e68-a541-8e53a4f5b862-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.725402 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/126e6e95-80d8-4e68-a541-8e53a4f5b862-config-data" (OuterVolumeSpecName: "config-data") pod "126e6e95-80d8-4e68-a541-8e53a4f5b862" (UID: "126e6e95-80d8-4e68-a541-8e53a4f5b862"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.743631 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/126e6e95-80d8-4e68-a541-8e53a4f5b862-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "126e6e95-80d8-4e68-a541-8e53a4f5b862" (UID: "126e6e95-80d8-4e68-a541-8e53a4f5b862"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.820234 4685 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/126e6e95-80d8-4e68-a541-8e53a4f5b862-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.820533 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/126e6e95-80d8-4e68-a541-8e53a4f5b862-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:33 crc kubenswrapper[4685]: I1013 09:02:33.850400 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 13 09:02:34 crc kubenswrapper[4685]: I1013 09:02:34.251033 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-n8774" event={"ID":"3688036a-1c46-45c2-ac9b-f2451a738f17","Type":"ContainerStarted","Data":"79eed5a1d46f13a5aed18173e0129a27427bd7ef0b550f6470c04a7136b3abdd"} Oct 13 09:02:34 crc kubenswrapper[4685]: I1013 09:02:34.251690 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:34 crc kubenswrapper[4685]: I1013 09:02:34.266413 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64d69c974d-22pqc" Oct 13 09:02:34 crc kubenswrapper[4685]: I1013 09:02:34.269780 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64d69c974d-22pqc" event={"ID":"126e6e95-80d8-4e68-a541-8e53a4f5b862","Type":"ContainerDied","Data":"34092f0c3725fb957baa150ccc4eaa3c17784f4348ba496fa3ee32ff339fac93"} Oct 13 09:02:34 crc kubenswrapper[4685]: I1013 09:02:34.269838 4685 scope.go:117] "RemoveContainer" containerID="b115917cd44da7ffaccbe91d120ee61d59a4460f53f619a7e3e0215ca2922259" Oct 13 09:02:34 crc kubenswrapper[4685]: I1013 09:02:34.273020 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-db976b79d-q446d" podUID="2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 13 09:02:34 crc kubenswrapper[4685]: I1013 09:02:34.273307 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/horizon-db976b79d-q446d" podUID="2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 13 09:02:34 crc kubenswrapper[4685]: I1013 09:02:34.277759 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-848cf88cfc-n8774" podStartSLOduration=11.277736728 podStartE2EDuration="11.277736728s" podCreationTimestamp="2025-10-13 09:02:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:02:34.276485034 +0000 UTC m=+1079.424360795" watchObservedRunningTime="2025-10-13 09:02:34.277736728 +0000 UTC m=+1079.425612489" Oct 13 09:02:34 crc kubenswrapper[4685]: I1013 09:02:34.311019 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-64d69c974d-22pqc"] Oct 13 09:02:34 crc kubenswrapper[4685]: I1013 09:02:34.316247 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-64d69c974d-22pqc"] Oct 13 09:02:34 crc kubenswrapper[4685]: I1013 09:02:34.713245 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 13 09:02:35 crc kubenswrapper[4685]: I1013 09:02:35.531040 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="126e6e95-80d8-4e68-a541-8e53a4f5b862" path="/var/lib/kubelet/pods/126e6e95-80d8-4e68-a541-8e53a4f5b862/volumes" Oct 13 09:02:36 crc kubenswrapper[4685]: I1013 09:02:36.539238 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:36 crc kubenswrapper[4685]: I1013 09:02:36.562346 4685 scope.go:117] "RemoveContainer" containerID="057d5f72232fe1033e30b206d78832bccc56b3093599e6d33773993e1467b3c1" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.102138 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.168027 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-55b747894d-xd2hx" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.207554 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-sg-core-conf-yaml\") pod \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.207596 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-combined-ca-bundle\") pod \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.207647 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-config-data\") pod \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.207715 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-scripts\") pod \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.207797 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4fhq\" (UniqueName: \"kubernetes.io/projected/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-kube-api-access-c4fhq\") pod \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.207823 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-run-httpd\") pod \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.207850 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-log-httpd\") pod \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\" (UID: \"0ffc408c-86d3-4e93-9d49-6ef39cb7d959\") " Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.208887 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0ffc408c-86d3-4e93-9d49-6ef39cb7d959" (UID: "0ffc408c-86d3-4e93-9d49-6ef39cb7d959"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.215758 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-scripts" (OuterVolumeSpecName: "scripts") pod "0ffc408c-86d3-4e93-9d49-6ef39cb7d959" (UID: "0ffc408c-86d3-4e93-9d49-6ef39cb7d959"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.238024 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0ffc408c-86d3-4e93-9d49-6ef39cb7d959" (UID: "0ffc408c-86d3-4e93-9d49-6ef39cb7d959"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.258446 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-7d4ff599b8-7thr6"] Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.258729 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-7d4ff599b8-7thr6" podUID="19f3def1-e242-41ec-8465-164d53d878a4" containerName="placement-log" containerID="cri-o://5ed93862142f8804adadb512e47f39fd79361493df4ab86de88674ec6691104b" gracePeriod=30 Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.264262 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-kube-api-access-c4fhq" (OuterVolumeSpecName: "kube-api-access-c4fhq") pod "0ffc408c-86d3-4e93-9d49-6ef39cb7d959" (UID: "0ffc408c-86d3-4e93-9d49-6ef39cb7d959"). InnerVolumeSpecName "kube-api-access-c4fhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.264575 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-7d4ff599b8-7thr6" podUID="19f3def1-e242-41ec-8465-164d53d878a4" containerName="placement-api" containerID="cri-o://3a44a77bb33a49bbaad3ced919a264013171d40d6bba7c0e36f5838e6c2ba239" gracePeriod=30 Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.285109 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0ffc408c-86d3-4e93-9d49-6ef39cb7d959" (UID: "0ffc408c-86d3-4e93-9d49-6ef39cb7d959"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.313645 4685 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.313678 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.313689 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4fhq\" (UniqueName: \"kubernetes.io/projected/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-kube-api-access-c4fhq\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.313699 4685 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.313707 4685 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.327789 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-f7b8c695-sw5hg" event={"ID":"102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11","Type":"ContainerStarted","Data":"eba1ff906f17d333302c4e963b7bb85d6990274f22df67e870551cbaf27771b4"} Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.333258 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-58b69dbf78-cznmg" event={"ID":"ad40cbfc-60f4-4ff0-9106-90a9941d5c10","Type":"ContainerStarted","Data":"2d38086fdd2df5cb428ae24b90372a3a7ac008ced5bf13033af53f48f796a45f"} Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.346062 4685 generic.go:334] "Generic (PLEG): container finished" podID="0ffc408c-86d3-4e93-9d49-6ef39cb7d959" containerID="da0a042d918652a2a33633572db233d9037398e55494df44006a3b65c0286815" exitCode=0 Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.347065 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.347592 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ffc408c-86d3-4e93-9d49-6ef39cb7d959","Type":"ContainerDied","Data":"da0a042d918652a2a33633572db233d9037398e55494df44006a3b65c0286815"} Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.348587 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ffc408c-86d3-4e93-9d49-6ef39cb7d959","Type":"ContainerDied","Data":"0306714f5cc052aeb656ce1d1a1366c895a35ae0f5259e364656562c360fa75e"} Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.348657 4685 scope.go:117] "RemoveContainer" containerID="39a2e5629a8f0cc1a200d4b9f131214c0abcf9eb6b5decfd5d264ac00f2ddfcb" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.401835 4685 scope.go:117] "RemoveContainer" containerID="dfd28dc5a0e2852cf335562850401d5db7a9d724c0154c5edec7099ac902d449" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.438628 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ffc408c-86d3-4e93-9d49-6ef39cb7d959" (UID: "0ffc408c-86d3-4e93-9d49-6ef39cb7d959"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.471223 4685 scope.go:117] "RemoveContainer" containerID="da0a042d918652a2a33633572db233d9037398e55494df44006a3b65c0286815" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.510698 4685 scope.go:117] "RemoveContainer" containerID="a392b4d4475d36fb966ec8ba4fcd6cb1bd6f904281b206bd0573f0b1765e6cf7" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.523857 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-config-data" (OuterVolumeSpecName: "config-data") pod "0ffc408c-86d3-4e93-9d49-6ef39cb7d959" (UID: "0ffc408c-86d3-4e93-9d49-6ef39cb7d959"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.525385 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.525422 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ffc408c-86d3-4e93-9d49-6ef39cb7d959-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.538349 4685 scope.go:117] "RemoveContainer" containerID="39a2e5629a8f0cc1a200d4b9f131214c0abcf9eb6b5decfd5d264ac00f2ddfcb" Oct 13 09:02:37 crc kubenswrapper[4685]: E1013 09:02:37.540601 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39a2e5629a8f0cc1a200d4b9f131214c0abcf9eb6b5decfd5d264ac00f2ddfcb\": container with ID starting with 39a2e5629a8f0cc1a200d4b9f131214c0abcf9eb6b5decfd5d264ac00f2ddfcb not found: ID does not exist" containerID="39a2e5629a8f0cc1a200d4b9f131214c0abcf9eb6b5decfd5d264ac00f2ddfcb" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.540634 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39a2e5629a8f0cc1a200d4b9f131214c0abcf9eb6b5decfd5d264ac00f2ddfcb"} err="failed to get container status \"39a2e5629a8f0cc1a200d4b9f131214c0abcf9eb6b5decfd5d264ac00f2ddfcb\": rpc error: code = NotFound desc = could not find container \"39a2e5629a8f0cc1a200d4b9f131214c0abcf9eb6b5decfd5d264ac00f2ddfcb\": container with ID starting with 39a2e5629a8f0cc1a200d4b9f131214c0abcf9eb6b5decfd5d264ac00f2ddfcb not found: ID does not exist" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.540654 4685 scope.go:117] "RemoveContainer" containerID="dfd28dc5a0e2852cf335562850401d5db7a9d724c0154c5edec7099ac902d449" Oct 13 09:02:37 crc kubenswrapper[4685]: E1013 09:02:37.541683 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfd28dc5a0e2852cf335562850401d5db7a9d724c0154c5edec7099ac902d449\": container with ID starting with dfd28dc5a0e2852cf335562850401d5db7a9d724c0154c5edec7099ac902d449 not found: ID does not exist" containerID="dfd28dc5a0e2852cf335562850401d5db7a9d724c0154c5edec7099ac902d449" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.541706 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfd28dc5a0e2852cf335562850401d5db7a9d724c0154c5edec7099ac902d449"} err="failed to get container status \"dfd28dc5a0e2852cf335562850401d5db7a9d724c0154c5edec7099ac902d449\": rpc error: code = NotFound desc = could not find container \"dfd28dc5a0e2852cf335562850401d5db7a9d724c0154c5edec7099ac902d449\": container with ID starting with dfd28dc5a0e2852cf335562850401d5db7a9d724c0154c5edec7099ac902d449 not found: ID does not exist" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.541719 4685 scope.go:117] "RemoveContainer" containerID="da0a042d918652a2a33633572db233d9037398e55494df44006a3b65c0286815" Oct 13 09:02:37 crc kubenswrapper[4685]: E1013 09:02:37.542094 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da0a042d918652a2a33633572db233d9037398e55494df44006a3b65c0286815\": container with ID starting with da0a042d918652a2a33633572db233d9037398e55494df44006a3b65c0286815 not found: ID does not exist" containerID="da0a042d918652a2a33633572db233d9037398e55494df44006a3b65c0286815" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.542182 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da0a042d918652a2a33633572db233d9037398e55494df44006a3b65c0286815"} err="failed to get container status \"da0a042d918652a2a33633572db233d9037398e55494df44006a3b65c0286815\": rpc error: code = NotFound desc = could not find container \"da0a042d918652a2a33633572db233d9037398e55494df44006a3b65c0286815\": container with ID starting with da0a042d918652a2a33633572db233d9037398e55494df44006a3b65c0286815 not found: ID does not exist" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.542195 4685 scope.go:117] "RemoveContainer" containerID="a392b4d4475d36fb966ec8ba4fcd6cb1bd6f904281b206bd0573f0b1765e6cf7" Oct 13 09:02:37 crc kubenswrapper[4685]: E1013 09:02:37.547121 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a392b4d4475d36fb966ec8ba4fcd6cb1bd6f904281b206bd0573f0b1765e6cf7\": container with ID starting with a392b4d4475d36fb966ec8ba4fcd6cb1bd6f904281b206bd0573f0b1765e6cf7 not found: ID does not exist" containerID="a392b4d4475d36fb966ec8ba4fcd6cb1bd6f904281b206bd0573f0b1765e6cf7" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.547175 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a392b4d4475d36fb966ec8ba4fcd6cb1bd6f904281b206bd0573f0b1765e6cf7"} err="failed to get container status \"a392b4d4475d36fb966ec8ba4fcd6cb1bd6f904281b206bd0573f0b1765e6cf7\": rpc error: code = NotFound desc = could not find container \"a392b4d4475d36fb966ec8ba4fcd6cb1bd6f904281b206bd0573f0b1765e6cf7\": container with ID starting with a392b4d4475d36fb966ec8ba4fcd6cb1bd6f904281b206bd0573f0b1765e6cf7 not found: ID does not exist" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.695294 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.706662 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.756644 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:02:37 crc kubenswrapper[4685]: E1013 09:02:37.757571 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ffc408c-86d3-4e93-9d49-6ef39cb7d959" containerName="sg-core" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.757592 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ffc408c-86d3-4e93-9d49-6ef39cb7d959" containerName="sg-core" Oct 13 09:02:37 crc kubenswrapper[4685]: E1013 09:02:37.757613 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ffc408c-86d3-4e93-9d49-6ef39cb7d959" containerName="proxy-httpd" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.757621 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ffc408c-86d3-4e93-9d49-6ef39cb7d959" containerName="proxy-httpd" Oct 13 09:02:37 crc kubenswrapper[4685]: E1013 09:02:37.757638 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ffc408c-86d3-4e93-9d49-6ef39cb7d959" containerName="ceilometer-central-agent" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.757644 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ffc408c-86d3-4e93-9d49-6ef39cb7d959" containerName="ceilometer-central-agent" Oct 13 09:02:37 crc kubenswrapper[4685]: E1013 09:02:37.757658 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="126e6e95-80d8-4e68-a541-8e53a4f5b862" containerName="horizon-log" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.757664 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="126e6e95-80d8-4e68-a541-8e53a4f5b862" containerName="horizon-log" Oct 13 09:02:37 crc kubenswrapper[4685]: E1013 09:02:37.757689 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ffc408c-86d3-4e93-9d49-6ef39cb7d959" containerName="ceilometer-notification-agent" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.757695 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ffc408c-86d3-4e93-9d49-6ef39cb7d959" containerName="ceilometer-notification-agent" Oct 13 09:02:37 crc kubenswrapper[4685]: E1013 09:02:37.757712 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c48e301-7038-4b0c-9bbf-dda3aa596fb4" containerName="dnsmasq-dns" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.757718 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c48e301-7038-4b0c-9bbf-dda3aa596fb4" containerName="dnsmasq-dns" Oct 13 09:02:37 crc kubenswrapper[4685]: E1013 09:02:37.757736 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="126e6e95-80d8-4e68-a541-8e53a4f5b862" containerName="horizon" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.757742 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="126e6e95-80d8-4e68-a541-8e53a4f5b862" containerName="horizon" Oct 13 09:02:37 crc kubenswrapper[4685]: E1013 09:02:37.757759 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c48e301-7038-4b0c-9bbf-dda3aa596fb4" containerName="init" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.757765 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c48e301-7038-4b0c-9bbf-dda3aa596fb4" containerName="init" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.758076 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ffc408c-86d3-4e93-9d49-6ef39cb7d959" containerName="proxy-httpd" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.758091 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="126e6e95-80d8-4e68-a541-8e53a4f5b862" containerName="horizon-log" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.758112 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c48e301-7038-4b0c-9bbf-dda3aa596fb4" containerName="dnsmasq-dns" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.758128 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ffc408c-86d3-4e93-9d49-6ef39cb7d959" containerName="ceilometer-notification-agent" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.758157 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="126e6e95-80d8-4e68-a541-8e53a4f5b862" containerName="horizon" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.758172 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ffc408c-86d3-4e93-9d49-6ef39cb7d959" containerName="ceilometer-central-agent" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.758208 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ffc408c-86d3-4e93-9d49-6ef39cb7d959" containerName="sg-core" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.769470 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.769627 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.776537 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.776875 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.938703 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-config-data\") pod \"ceilometer-0\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " pod="openstack/ceilometer-0" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.938799 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a6acbb5-144a-42d3-a770-60b0fe01de1f-log-httpd\") pod \"ceilometer-0\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " pod="openstack/ceilometer-0" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.938835 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " pod="openstack/ceilometer-0" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.938851 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a6acbb5-144a-42d3-a770-60b0fe01de1f-run-httpd\") pod \"ceilometer-0\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " pod="openstack/ceilometer-0" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.939003 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " pod="openstack/ceilometer-0" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.939057 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwthr\" (UniqueName: \"kubernetes.io/projected/7a6acbb5-144a-42d3-a770-60b0fe01de1f-kube-api-access-bwthr\") pod \"ceilometer-0\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " pod="openstack/ceilometer-0" Oct 13 09:02:37 crc kubenswrapper[4685]: I1013 09:02:37.939103 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-scripts\") pod \"ceilometer-0\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " pod="openstack/ceilometer-0" Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.040521 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " pod="openstack/ceilometer-0" Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.040567 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwthr\" (UniqueName: \"kubernetes.io/projected/7a6acbb5-144a-42d3-a770-60b0fe01de1f-kube-api-access-bwthr\") pod \"ceilometer-0\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " pod="openstack/ceilometer-0" Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.040600 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-scripts\") pod \"ceilometer-0\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " pod="openstack/ceilometer-0" Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.040650 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-config-data\") pod \"ceilometer-0\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " pod="openstack/ceilometer-0" Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.040704 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a6acbb5-144a-42d3-a770-60b0fe01de1f-log-httpd\") pod \"ceilometer-0\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " pod="openstack/ceilometer-0" Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.040737 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " pod="openstack/ceilometer-0" Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.040752 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a6acbb5-144a-42d3-a770-60b0fe01de1f-run-httpd\") pod \"ceilometer-0\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " pod="openstack/ceilometer-0" Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.041140 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a6acbb5-144a-42d3-a770-60b0fe01de1f-run-httpd\") pod \"ceilometer-0\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " pod="openstack/ceilometer-0" Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.042127 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a6acbb5-144a-42d3-a770-60b0fe01de1f-log-httpd\") pod \"ceilometer-0\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " pod="openstack/ceilometer-0" Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.047336 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-scripts\") pod \"ceilometer-0\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " pod="openstack/ceilometer-0" Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.056903 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " pod="openstack/ceilometer-0" Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.057738 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " pod="openstack/ceilometer-0" Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.058533 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwthr\" (UniqueName: \"kubernetes.io/projected/7a6acbb5-144a-42d3-a770-60b0fe01de1f-kube-api-access-bwthr\") pod \"ceilometer-0\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " pod="openstack/ceilometer-0" Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.062007 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-config-data\") pod \"ceilometer-0\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " pod="openstack/ceilometer-0" Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.113611 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.372092 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-f7b8c695-sw5hg" event={"ID":"102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11","Type":"ContainerStarted","Data":"b6fd4d59e82f114812ba8ac3fa0ada3280b8858d2da8d84eb6f2cfed3dd3636d"} Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.388076 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-58b69dbf78-cznmg" event={"ID":"ad40cbfc-60f4-4ff0-9106-90a9941d5c10","Type":"ContainerStarted","Data":"32b768a504aeb7598f869fde730a05de81c25042a83204a3ac07bd79ceaa657b"} Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.392669 4685 generic.go:334] "Generic (PLEG): container finished" podID="19f3def1-e242-41ec-8465-164d53d878a4" containerID="5ed93862142f8804adadb512e47f39fd79361493df4ab86de88674ec6691104b" exitCode=143 Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.392716 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d4ff599b8-7thr6" event={"ID":"19f3def1-e242-41ec-8465-164d53d878a4","Type":"ContainerDied","Data":"5ed93862142f8804adadb512e47f39fd79361493df4ab86de88674ec6691104b"} Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.406321 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-f7b8c695-sw5hg" podStartSLOduration=11.182580095 podStartE2EDuration="16.406304491s" podCreationTimestamp="2025-10-13 09:02:22 +0000 UTC" firstStartedPulling="2025-10-13 09:02:31.438268092 +0000 UTC m=+1076.586143853" lastFinishedPulling="2025-10-13 09:02:36.661992498 +0000 UTC m=+1081.809868249" observedRunningTime="2025-10-13 09:02:38.394220702 +0000 UTC m=+1083.542096463" watchObservedRunningTime="2025-10-13 09:02:38.406304491 +0000 UTC m=+1083.554180252" Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.423421 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-58b69dbf78-cznmg" podStartSLOduration=11.159353611 podStartE2EDuration="16.423406368s" podCreationTimestamp="2025-10-13 09:02:22 +0000 UTC" firstStartedPulling="2025-10-13 09:02:31.394416735 +0000 UTC m=+1076.542292496" lastFinishedPulling="2025-10-13 09:02:36.658469492 +0000 UTC m=+1081.806345253" observedRunningTime="2025-10-13 09:02:38.419340387 +0000 UTC m=+1083.567216148" watchObservedRunningTime="2025-10-13 09:02:38.423406368 +0000 UTC m=+1083.571282129" Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.435087 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.494289 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-8hg5b"] Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.494534 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" podUID="778d19c2-2541-40d9-a00c-efb46a746a32" containerName="dnsmasq-dns" containerID="cri-o://1ee88d52907723bc1c8a7413e7086126a5b37a8ce4fa2dceb8104a586c54fac2" gracePeriod=10 Oct 13 09:02:38 crc kubenswrapper[4685]: I1013 09:02:38.626552 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.073468 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.167577 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-ovsdbserver-sb\") pod \"778d19c2-2541-40d9-a00c-efb46a746a32\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.167675 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-ovsdbserver-nb\") pod \"778d19c2-2541-40d9-a00c-efb46a746a32\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.167721 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-config\") pod \"778d19c2-2541-40d9-a00c-efb46a746a32\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.167748 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-dns-svc\") pod \"778d19c2-2541-40d9-a00c-efb46a746a32\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.167811 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-dns-swift-storage-0\") pod \"778d19c2-2541-40d9-a00c-efb46a746a32\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.167841 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c5x2\" (UniqueName: \"kubernetes.io/projected/778d19c2-2541-40d9-a00c-efb46a746a32-kube-api-access-7c5x2\") pod \"778d19c2-2541-40d9-a00c-efb46a746a32\" (UID: \"778d19c2-2541-40d9-a00c-efb46a746a32\") " Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.200214 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/778d19c2-2541-40d9-a00c-efb46a746a32-kube-api-access-7c5x2" (OuterVolumeSpecName: "kube-api-access-7c5x2") pod "778d19c2-2541-40d9-a00c-efb46a746a32" (UID: "778d19c2-2541-40d9-a00c-efb46a746a32"). InnerVolumeSpecName "kube-api-access-7c5x2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.269903 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c5x2\" (UniqueName: \"kubernetes.io/projected/778d19c2-2541-40d9-a00c-efb46a746a32-kube-api-access-7c5x2\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.291677 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "778d19c2-2541-40d9-a00c-efb46a746a32" (UID: "778d19c2-2541-40d9-a00c-efb46a746a32"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.331427 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "778d19c2-2541-40d9-a00c-efb46a746a32" (UID: "778d19c2-2541-40d9-a00c-efb46a746a32"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.347349 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "778d19c2-2541-40d9-a00c-efb46a746a32" (UID: "778d19c2-2541-40d9-a00c-efb46a746a32"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.364393 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-config" (OuterVolumeSpecName: "config") pod "778d19c2-2541-40d9-a00c-efb46a746a32" (UID: "778d19c2-2541-40d9-a00c-efb46a746a32"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.371996 4685 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.372022 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.372035 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-config\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.372043 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.380433 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "778d19c2-2541-40d9-a00c-efb46a746a32" (UID: "778d19c2-2541-40d9-a00c-efb46a746a32"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.406403 4685 generic.go:334] "Generic (PLEG): container finished" podID="778d19c2-2541-40d9-a00c-efb46a746a32" containerID="1ee88d52907723bc1c8a7413e7086126a5b37a8ce4fa2dceb8104a586c54fac2" exitCode=0 Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.406458 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" event={"ID":"778d19c2-2541-40d9-a00c-efb46a746a32","Type":"ContainerDied","Data":"1ee88d52907723bc1c8a7413e7086126a5b37a8ce4fa2dceb8104a586c54fac2"} Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.406483 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" event={"ID":"778d19c2-2541-40d9-a00c-efb46a746a32","Type":"ContainerDied","Data":"d5f0bc39baa4772dd60c30384357f016b1c146d66cf132db7e67931f44542c0a"} Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.406498 4685 scope.go:117] "RemoveContainer" containerID="1ee88d52907723bc1c8a7413e7086126a5b37a8ce4fa2dceb8104a586c54fac2" Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.406627 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-8hg5b" Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.415845 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a6acbb5-144a-42d3-a770-60b0fe01de1f","Type":"ContainerStarted","Data":"3215d07ac65a0b9fbfc92d29b679c8b97f42551c1674456f66bea44f94feaa6a"} Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.439020 4685 scope.go:117] "RemoveContainer" containerID="8fbbf169b94346ba341c373b92ee4c69f7cf5cb3eeaf4f184ceba78ab8924481" Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.444560 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-8hg5b"] Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.456398 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-8hg5b"] Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.474306 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/778d19c2-2541-40d9-a00c-efb46a746a32-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.494238 4685 scope.go:117] "RemoveContainer" containerID="1ee88d52907723bc1c8a7413e7086126a5b37a8ce4fa2dceb8104a586c54fac2" Oct 13 09:02:39 crc kubenswrapper[4685]: E1013 09:02:39.494809 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ee88d52907723bc1c8a7413e7086126a5b37a8ce4fa2dceb8104a586c54fac2\": container with ID starting with 1ee88d52907723bc1c8a7413e7086126a5b37a8ce4fa2dceb8104a586c54fac2 not found: ID does not exist" containerID="1ee88d52907723bc1c8a7413e7086126a5b37a8ce4fa2dceb8104a586c54fac2" Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.494838 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ee88d52907723bc1c8a7413e7086126a5b37a8ce4fa2dceb8104a586c54fac2"} err="failed to get container status \"1ee88d52907723bc1c8a7413e7086126a5b37a8ce4fa2dceb8104a586c54fac2\": rpc error: code = NotFound desc = could not find container \"1ee88d52907723bc1c8a7413e7086126a5b37a8ce4fa2dceb8104a586c54fac2\": container with ID starting with 1ee88d52907723bc1c8a7413e7086126a5b37a8ce4fa2dceb8104a586c54fac2 not found: ID does not exist" Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.494856 4685 scope.go:117] "RemoveContainer" containerID="8fbbf169b94346ba341c373b92ee4c69f7cf5cb3eeaf4f184ceba78ab8924481" Oct 13 09:02:39 crc kubenswrapper[4685]: E1013 09:02:39.495868 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fbbf169b94346ba341c373b92ee4c69f7cf5cb3eeaf4f184ceba78ab8924481\": container with ID starting with 8fbbf169b94346ba341c373b92ee4c69f7cf5cb3eeaf4f184ceba78ab8924481 not found: ID does not exist" containerID="8fbbf169b94346ba341c373b92ee4c69f7cf5cb3eeaf4f184ceba78ab8924481" Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.495889 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fbbf169b94346ba341c373b92ee4c69f7cf5cb3eeaf4f184ceba78ab8924481"} err="failed to get container status \"8fbbf169b94346ba341c373b92ee4c69f7cf5cb3eeaf4f184ceba78ab8924481\": rpc error: code = NotFound desc = could not find container \"8fbbf169b94346ba341c373b92ee4c69f7cf5cb3eeaf4f184ceba78ab8924481\": container with ID starting with 8fbbf169b94346ba341c373b92ee4c69f7cf5cb3eeaf4f184ceba78ab8924481 not found: ID does not exist" Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.524729 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ffc408c-86d3-4e93-9d49-6ef39cb7d959" path="/var/lib/kubelet/pods/0ffc408c-86d3-4e93-9d49-6ef39cb7d959/volumes" Oct 13 09:02:39 crc kubenswrapper[4685]: I1013 09:02:39.526423 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="778d19c2-2541-40d9-a00c-efb46a746a32" path="/var/lib/kubelet/pods/778d19c2-2541-40d9-a00c-efb46a746a32/volumes" Oct 13 09:02:40 crc kubenswrapper[4685]: I1013 09:02:40.429456 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a6acbb5-144a-42d3-a770-60b0fe01de1f","Type":"ContainerStarted","Data":"d57b8e8b608b05b8d83161391dfa3b7f3461bc38b72ef48072d5ef4d5b6a8352"} Oct 13 09:02:40 crc kubenswrapper[4685]: I1013 09:02:40.966411 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.145876 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.320819 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-config-data\") pod \"19f3def1-e242-41ec-8465-164d53d878a4\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.321285 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-combined-ca-bundle\") pod \"19f3def1-e242-41ec-8465-164d53d878a4\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.321364 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sq75x\" (UniqueName: \"kubernetes.io/projected/19f3def1-e242-41ec-8465-164d53d878a4-kube-api-access-sq75x\") pod \"19f3def1-e242-41ec-8465-164d53d878a4\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.321397 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19f3def1-e242-41ec-8465-164d53d878a4-logs\") pod \"19f3def1-e242-41ec-8465-164d53d878a4\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.321444 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-internal-tls-certs\") pod \"19f3def1-e242-41ec-8465-164d53d878a4\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.321465 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-scripts\") pod \"19f3def1-e242-41ec-8465-164d53d878a4\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.321494 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-public-tls-certs\") pod \"19f3def1-e242-41ec-8465-164d53d878a4\" (UID: \"19f3def1-e242-41ec-8465-164d53d878a4\") " Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.322294 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19f3def1-e242-41ec-8465-164d53d878a4-logs" (OuterVolumeSpecName: "logs") pod "19f3def1-e242-41ec-8465-164d53d878a4" (UID: "19f3def1-e242-41ec-8465-164d53d878a4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.342385 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19f3def1-e242-41ec-8465-164d53d878a4-kube-api-access-sq75x" (OuterVolumeSpecName: "kube-api-access-sq75x") pod "19f3def1-e242-41ec-8465-164d53d878a4" (UID: "19f3def1-e242-41ec-8465-164d53d878a4"). InnerVolumeSpecName "kube-api-access-sq75x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.345440 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-scripts" (OuterVolumeSpecName: "scripts") pod "19f3def1-e242-41ec-8465-164d53d878a4" (UID: "19f3def1-e242-41ec-8465-164d53d878a4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.424292 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sq75x\" (UniqueName: \"kubernetes.io/projected/19f3def1-e242-41ec-8465-164d53d878a4-kube-api-access-sq75x\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.424338 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19f3def1-e242-41ec-8465-164d53d878a4-logs\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.424349 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.428891 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-config-data" (OuterVolumeSpecName: "config-data") pod "19f3def1-e242-41ec-8465-164d53d878a4" (UID: "19f3def1-e242-41ec-8465-164d53d878a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.455946 4685 generic.go:334] "Generic (PLEG): container finished" podID="19f3def1-e242-41ec-8465-164d53d878a4" containerID="3a44a77bb33a49bbaad3ced919a264013171d40d6bba7c0e36f5838e6c2ba239" exitCode=0 Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.456025 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d4ff599b8-7thr6" event={"ID":"19f3def1-e242-41ec-8465-164d53d878a4","Type":"ContainerDied","Data":"3a44a77bb33a49bbaad3ced919a264013171d40d6bba7c0e36f5838e6c2ba239"} Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.456053 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d4ff599b8-7thr6" event={"ID":"19f3def1-e242-41ec-8465-164d53d878a4","Type":"ContainerDied","Data":"2ceb06eecac6b5172b62220dc937acd208d280a46058c9d9b0cb2269ce2785ed"} Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.456094 4685 scope.go:117] "RemoveContainer" containerID="3a44a77bb33a49bbaad3ced919a264013171d40d6bba7c0e36f5838e6c2ba239" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.456275 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7d4ff599b8-7thr6" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.460831 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a6acbb5-144a-42d3-a770-60b0fe01de1f","Type":"ContainerStarted","Data":"0824af0ad0d3226e908ed1d3fa419c6cb802a46b76ee0af084c4f6a3adfffe2f"} Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.496476 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19f3def1-e242-41ec-8465-164d53d878a4" (UID: "19f3def1-e242-41ec-8465-164d53d878a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.496508 4685 scope.go:117] "RemoveContainer" containerID="5ed93862142f8804adadb512e47f39fd79361493df4ab86de88674ec6691104b" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.511535 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "19f3def1-e242-41ec-8465-164d53d878a4" (UID: "19f3def1-e242-41ec-8465-164d53d878a4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.518147 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "19f3def1-e242-41ec-8465-164d53d878a4" (UID: "19f3def1-e242-41ec-8465-164d53d878a4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.519428 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.525653 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.525682 4685 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.525693 4685 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.525703 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19f3def1-e242-41ec-8465-164d53d878a4-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.541625 4685 scope.go:117] "RemoveContainer" containerID="3a44a77bb33a49bbaad3ced919a264013171d40d6bba7c0e36f5838e6c2ba239" Oct 13 09:02:41 crc kubenswrapper[4685]: E1013 09:02:41.542496 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a44a77bb33a49bbaad3ced919a264013171d40d6bba7c0e36f5838e6c2ba239\": container with ID starting with 3a44a77bb33a49bbaad3ced919a264013171d40d6bba7c0e36f5838e6c2ba239 not found: ID does not exist" containerID="3a44a77bb33a49bbaad3ced919a264013171d40d6bba7c0e36f5838e6c2ba239" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.542530 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a44a77bb33a49bbaad3ced919a264013171d40d6bba7c0e36f5838e6c2ba239"} err="failed to get container status \"3a44a77bb33a49bbaad3ced919a264013171d40d6bba7c0e36f5838e6c2ba239\": rpc error: code = NotFound desc = could not find container \"3a44a77bb33a49bbaad3ced919a264013171d40d6bba7c0e36f5838e6c2ba239\": container with ID starting with 3a44a77bb33a49bbaad3ced919a264013171d40d6bba7c0e36f5838e6c2ba239 not found: ID does not exist" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.542562 4685 scope.go:117] "RemoveContainer" containerID="5ed93862142f8804adadb512e47f39fd79361493df4ab86de88674ec6691104b" Oct 13 09:02:41 crc kubenswrapper[4685]: E1013 09:02:41.543035 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ed93862142f8804adadb512e47f39fd79361493df4ab86de88674ec6691104b\": container with ID starting with 5ed93862142f8804adadb512e47f39fd79361493df4ab86de88674ec6691104b not found: ID does not exist" containerID="5ed93862142f8804adadb512e47f39fd79361493df4ab86de88674ec6691104b" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.543058 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ed93862142f8804adadb512e47f39fd79361493df4ab86de88674ec6691104b"} err="failed to get container status \"5ed93862142f8804adadb512e47f39fd79361493df4ab86de88674ec6691104b\": rpc error: code = NotFound desc = could not find container \"5ed93862142f8804adadb512e47f39fd79361493df4ab86de88674ec6691104b\": container with ID starting with 5ed93862142f8804adadb512e47f39fd79361493df4ab86de88674ec6691104b not found: ID does not exist" Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.785830 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-7d4ff599b8-7thr6"] Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.795793 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-7d4ff599b8-7thr6"] Oct 13 09:02:41 crc kubenswrapper[4685]: I1013 09:02:41.968673 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:43 crc kubenswrapper[4685]: I1013 09:02:43.056068 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-dbcf576b-msgft" podUID="46272601-40bd-43ec-b7a8-6824281972e0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Oct 13 09:02:43 crc kubenswrapper[4685]: I1013 09:02:43.454750 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7cc9c86456-5hsl7" Oct 13 09:02:43 crc kubenswrapper[4685]: I1013 09:02:43.486694 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a6acbb5-144a-42d3-a770-60b0fe01de1f","Type":"ContainerStarted","Data":"da414155495a5e007b148dee7722c2f3ca121974df04632a00cae29c3944bfa4"} Oct 13 09:02:43 crc kubenswrapper[4685]: I1013 09:02:43.488238 4685 generic.go:334] "Generic (PLEG): container finished" podID="b225d240-40a0-4d59-aa19-6355515e9f8c" containerID="048241ab2b6e54a12029b77e838f010d8527358e132afb5585f66a15de926667" exitCode=0 Oct 13 09:02:43 crc kubenswrapper[4685]: I1013 09:02:43.488280 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7h6qb" event={"ID":"b225d240-40a0-4d59-aa19-6355515e9f8c","Type":"ContainerDied","Data":"048241ab2b6e54a12029b77e838f010d8527358e132afb5585f66a15de926667"} Oct 13 09:02:43 crc kubenswrapper[4685]: I1013 09:02:43.525424 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19f3def1-e242-41ec-8465-164d53d878a4" path="/var/lib/kubelet/pods/19f3def1-e242-41ec-8465-164d53d878a4/volumes" Oct 13 09:02:43 crc kubenswrapper[4685]: I1013 09:02:43.526000 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5746d5f994-pjpl9"] Oct 13 09:02:43 crc kubenswrapper[4685]: I1013 09:02:43.526168 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5746d5f994-pjpl9" podUID="9dd392f6-cf1c-4d69-81f8-167718c41926" containerName="barbican-api-log" containerID="cri-o://cbe6d0a2722732a4d111f053aa7a593a39474ddfd32b4f71164834811c6471d1" gracePeriod=30 Oct 13 09:02:43 crc kubenswrapper[4685]: I1013 09:02:43.526410 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5746d5f994-pjpl9" podUID="9dd392f6-cf1c-4d69-81f8-167718c41926" containerName="barbican-api" containerID="cri-o://b559fba90dced4d180fb0896393fea41366c1de9c4acd2e31f5bf95bd3823d3b" gracePeriod=30 Oct 13 09:02:43 crc kubenswrapper[4685]: I1013 09:02:43.537779 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5746d5f994-pjpl9" podUID="9dd392f6-cf1c-4d69-81f8-167718c41926" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": EOF" Oct 13 09:02:43 crc kubenswrapper[4685]: I1013 09:02:43.537843 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5746d5f994-pjpl9" podUID="9dd392f6-cf1c-4d69-81f8-167718c41926" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": EOF" Oct 13 09:02:43 crc kubenswrapper[4685]: I1013 09:02:43.537847 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5746d5f994-pjpl9" podUID="9dd392f6-cf1c-4d69-81f8-167718c41926" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": EOF" Oct 13 09:02:44 crc kubenswrapper[4685]: I1013 09:02:44.498786 4685 generic.go:334] "Generic (PLEG): container finished" podID="9dd392f6-cf1c-4d69-81f8-167718c41926" containerID="cbe6d0a2722732a4d111f053aa7a593a39474ddfd32b4f71164834811c6471d1" exitCode=143 Oct 13 09:02:44 crc kubenswrapper[4685]: I1013 09:02:44.498867 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5746d5f994-pjpl9" event={"ID":"9dd392f6-cf1c-4d69-81f8-167718c41926","Type":"ContainerDied","Data":"cbe6d0a2722732a4d111f053aa7a593a39474ddfd32b4f71164834811c6471d1"} Oct 13 09:02:44 crc kubenswrapper[4685]: I1013 09:02:44.919355 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7h6qb" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.012441 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmq9p\" (UniqueName: \"kubernetes.io/projected/b225d240-40a0-4d59-aa19-6355515e9f8c-kube-api-access-fmq9p\") pod \"b225d240-40a0-4d59-aa19-6355515e9f8c\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.012772 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-db-sync-config-data\") pod \"b225d240-40a0-4d59-aa19-6355515e9f8c\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.012813 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-scripts\") pod \"b225d240-40a0-4d59-aa19-6355515e9f8c\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.012837 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-combined-ca-bundle\") pod \"b225d240-40a0-4d59-aa19-6355515e9f8c\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.012885 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-config-data\") pod \"b225d240-40a0-4d59-aa19-6355515e9f8c\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.012943 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b225d240-40a0-4d59-aa19-6355515e9f8c-etc-machine-id\") pod \"b225d240-40a0-4d59-aa19-6355515e9f8c\" (UID: \"b225d240-40a0-4d59-aa19-6355515e9f8c\") " Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.013248 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b225d240-40a0-4d59-aa19-6355515e9f8c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b225d240-40a0-4d59-aa19-6355515e9f8c" (UID: "b225d240-40a0-4d59-aa19-6355515e9f8c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.017291 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-scripts" (OuterVolumeSpecName: "scripts") pod "b225d240-40a0-4d59-aa19-6355515e9f8c" (UID: "b225d240-40a0-4d59-aa19-6355515e9f8c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.018577 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b225d240-40a0-4d59-aa19-6355515e9f8c" (UID: "b225d240-40a0-4d59-aa19-6355515e9f8c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.033226 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b225d240-40a0-4d59-aa19-6355515e9f8c-kube-api-access-fmq9p" (OuterVolumeSpecName: "kube-api-access-fmq9p") pod "b225d240-40a0-4d59-aa19-6355515e9f8c" (UID: "b225d240-40a0-4d59-aa19-6355515e9f8c"). InnerVolumeSpecName "kube-api-access-fmq9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.056990 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b225d240-40a0-4d59-aa19-6355515e9f8c" (UID: "b225d240-40a0-4d59-aa19-6355515e9f8c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.105886 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-config-data" (OuterVolumeSpecName: "config-data") pod "b225d240-40a0-4d59-aa19-6355515e9f8c" (UID: "b225d240-40a0-4d59-aa19-6355515e9f8c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.114500 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmq9p\" (UniqueName: \"kubernetes.io/projected/b225d240-40a0-4d59-aa19-6355515e9f8c-kube-api-access-fmq9p\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.114530 4685 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.114539 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.114550 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.114561 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b225d240-40a0-4d59-aa19-6355515e9f8c-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.114572 4685 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b225d240-40a0-4d59-aa19-6355515e9f8c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.527278 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a6acbb5-144a-42d3-a770-60b0fe01de1f","Type":"ContainerStarted","Data":"ceb48dda53ea7f8d1d12ae899581138421aa4e83c663333e32936fb5902d73af"} Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.528087 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.529395 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7h6qb" event={"ID":"b225d240-40a0-4d59-aa19-6355515e9f8c","Type":"ContainerDied","Data":"6c9cd308010bd180f39154e0de4b0369c02187761d7e1ecfa996b4b513a87b6b"} Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.529418 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c9cd308010bd180f39154e0de4b0369c02187761d7e1ecfa996b4b513a87b6b" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.529454 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7h6qb" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.578796 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.245720394 podStartE2EDuration="8.57877819s" podCreationTimestamp="2025-10-13 09:02:37 +0000 UTC" firstStartedPulling="2025-10-13 09:02:38.66417134 +0000 UTC m=+1083.812047101" lastFinishedPulling="2025-10-13 09:02:44.997229116 +0000 UTC m=+1090.145104897" observedRunningTime="2025-10-13 09:02:45.571595654 +0000 UTC m=+1090.719471415" watchObservedRunningTime="2025-10-13 09:02:45.57877819 +0000 UTC m=+1090.726653951" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.893068 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 13 09:02:45 crc kubenswrapper[4685]: E1013 09:02:45.893801 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="778d19c2-2541-40d9-a00c-efb46a746a32" containerName="dnsmasq-dns" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.893823 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="778d19c2-2541-40d9-a00c-efb46a746a32" containerName="dnsmasq-dns" Oct 13 09:02:45 crc kubenswrapper[4685]: E1013 09:02:45.893847 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b225d240-40a0-4d59-aa19-6355515e9f8c" containerName="cinder-db-sync" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.893854 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b225d240-40a0-4d59-aa19-6355515e9f8c" containerName="cinder-db-sync" Oct 13 09:02:45 crc kubenswrapper[4685]: E1013 09:02:45.893869 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19f3def1-e242-41ec-8465-164d53d878a4" containerName="placement-api" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.893876 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="19f3def1-e242-41ec-8465-164d53d878a4" containerName="placement-api" Oct 13 09:02:45 crc kubenswrapper[4685]: E1013 09:02:45.893910 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="778d19c2-2541-40d9-a00c-efb46a746a32" containerName="init" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.893934 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="778d19c2-2541-40d9-a00c-efb46a746a32" containerName="init" Oct 13 09:02:45 crc kubenswrapper[4685]: E1013 09:02:45.893954 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19f3def1-e242-41ec-8465-164d53d878a4" containerName="placement-log" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.893962 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="19f3def1-e242-41ec-8465-164d53d878a4" containerName="placement-log" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.894156 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="778d19c2-2541-40d9-a00c-efb46a746a32" containerName="dnsmasq-dns" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.894182 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b225d240-40a0-4d59-aa19-6355515e9f8c" containerName="cinder-db-sync" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.894209 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="19f3def1-e242-41ec-8465-164d53d878a4" containerName="placement-api" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.894225 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="19f3def1-e242-41ec-8465-164d53d878a4" containerName="placement-log" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.895180 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.903206 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-9966b" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.903523 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.906462 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.907087 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 13 09:02:45 crc kubenswrapper[4685]: I1013 09:02:45.992779 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.033272 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-7c4gj"] Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.035806 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.035892 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-scripts\") pod \"cinder-scheduler-0\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.036017 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-config-data\") pod \"cinder-scheduler-0\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.036103 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.036152 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.036176 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl84c\" (UniqueName: \"kubernetes.io/projected/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-kube-api-access-vl84c\") pod \"cinder-scheduler-0\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.036204 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.065238 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-7c4gj"] Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.142743 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-scripts\") pod \"cinder-scheduler-0\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.142889 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-config-data\") pod \"cinder-scheduler-0\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.142974 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-dns-svc\") pod \"dnsmasq-dns-6578955fd5-7c4gj\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.142990 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-7c4gj\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.143007 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-7c4gj\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.143094 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.143140 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb6kz\" (UniqueName: \"kubernetes.io/projected/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-kube-api-access-mb6kz\") pod \"dnsmasq-dns-6578955fd5-7c4gj\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.143167 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-config\") pod \"dnsmasq-dns-6578955fd5-7c4gj\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.143186 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.143204 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl84c\" (UniqueName: \"kubernetes.io/projected/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-kube-api-access-vl84c\") pod \"cinder-scheduler-0\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.143231 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.143255 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-7c4gj\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.143590 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.148745 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.149667 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.170372 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-scripts\") pod \"cinder-scheduler-0\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.171384 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-config-data\") pod \"cinder-scheduler-0\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.194152 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl84c\" (UniqueName: \"kubernetes.io/projected/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-kube-api-access-vl84c\") pod \"cinder-scheduler-0\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.207060 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.208779 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.216457 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.224224 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.224613 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.245253 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-dns-svc\") pod \"dnsmasq-dns-6578955fd5-7c4gj\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.248443 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-dns-svc\") pod \"dnsmasq-dns-6578955fd5-7c4gj\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.249492 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-7c4gj\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.249574 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-7c4gj\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.249690 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb6kz\" (UniqueName: \"kubernetes.io/projected/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-kube-api-access-mb6kz\") pod \"dnsmasq-dns-6578955fd5-7c4gj\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.249721 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-config\") pod \"dnsmasq-dns-6578955fd5-7c4gj\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.249759 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-7c4gj\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.250499 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-7c4gj\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.251051 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-7c4gj\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.251748 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-config\") pod \"dnsmasq-dns-6578955fd5-7c4gj\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.252181 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-7c4gj\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.272849 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb6kz\" (UniqueName: \"kubernetes.io/projected/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-kube-api-access-mb6kz\") pod \"dnsmasq-dns-6578955fd5-7c4gj\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.351726 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxpt4\" (UniqueName: \"kubernetes.io/projected/548142e8-94bd-4714-b65d-a4ec695000df-kube-api-access-bxpt4\") pod \"cinder-api-0\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.351806 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-scripts\") pod \"cinder-api-0\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.351826 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-config-data\") pod \"cinder-api-0\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.351843 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/548142e8-94bd-4714-b65d-a4ec695000df-logs\") pod \"cinder-api-0\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.351862 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/548142e8-94bd-4714-b65d-a4ec695000df-etc-machine-id\") pod \"cinder-api-0\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.351889 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.351934 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-config-data-custom\") pod \"cinder-api-0\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.398002 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.454869 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-config-data-custom\") pod \"cinder-api-0\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.455033 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxpt4\" (UniqueName: \"kubernetes.io/projected/548142e8-94bd-4714-b65d-a4ec695000df-kube-api-access-bxpt4\") pod \"cinder-api-0\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.455093 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-scripts\") pod \"cinder-api-0\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.455112 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-config-data\") pod \"cinder-api-0\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.455126 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/548142e8-94bd-4714-b65d-a4ec695000df-logs\") pod \"cinder-api-0\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.455143 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/548142e8-94bd-4714-b65d-a4ec695000df-etc-machine-id\") pod \"cinder-api-0\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.455166 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.462454 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/548142e8-94bd-4714-b65d-a4ec695000df-logs\") pod \"cinder-api-0\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.468618 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.468689 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/548142e8-94bd-4714-b65d-a4ec695000df-etc-machine-id\") pod \"cinder-api-0\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.469433 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-scripts\") pod \"cinder-api-0\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.477134 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-config-data\") pod \"cinder-api-0\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.477675 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-config-data-custom\") pod \"cinder-api-0\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.504400 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxpt4\" (UniqueName: \"kubernetes.io/projected/548142e8-94bd-4714-b65d-a4ec695000df-kube-api-access-bxpt4\") pod \"cinder-api-0\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.640692 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 13 09:02:46 crc kubenswrapper[4685]: I1013 09:02:46.961531 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 13 09:02:47 crc kubenswrapper[4685]: W1013 09:02:47.146096 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a253ee4_12d7_4e32_9d53_27ecc16ce8cb.slice/crio-f755cd6e41eba77fc76e07d259f1af977e371a9faa7beb1a67f52ca4ead82811 WatchSource:0}: Error finding container f755cd6e41eba77fc76e07d259f1af977e371a9faa7beb1a67f52ca4ead82811: Status 404 returned error can't find the container with id f755cd6e41eba77fc76e07d259f1af977e371a9faa7beb1a67f52ca4ead82811 Oct 13 09:02:47 crc kubenswrapper[4685]: I1013 09:02:47.150081 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-7c4gj"] Oct 13 09:02:47 crc kubenswrapper[4685]: I1013 09:02:47.291557 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 13 09:02:47 crc kubenswrapper[4685]: W1013 09:02:47.297768 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod548142e8_94bd_4714_b65d_a4ec695000df.slice/crio-c4f0a3d7ac361877440179c47d193b291be237e575dd0029c1520c931dd1a135 WatchSource:0}: Error finding container c4f0a3d7ac361877440179c47d193b291be237e575dd0029c1520c931dd1a135: Status 404 returned error can't find the container with id c4f0a3d7ac361877440179c47d193b291be237e575dd0029c1520c931dd1a135 Oct 13 09:02:47 crc kubenswrapper[4685]: I1013 09:02:47.466231 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5c8fd5c86b-qkv7x" Oct 13 09:02:47 crc kubenswrapper[4685]: I1013 09:02:47.604897 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1bfd68fe-249f-43a2-8991-dbbe8eb5e184","Type":"ContainerStarted","Data":"e69d534659382dc4a7c4f1eb67bfcbf8e9de42febc83cf1a11a29c2068841ea9"} Oct 13 09:02:47 crc kubenswrapper[4685]: I1013 09:02:47.616314 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"548142e8-94bd-4714-b65d-a4ec695000df","Type":"ContainerStarted","Data":"c4f0a3d7ac361877440179c47d193b291be237e575dd0029c1520c931dd1a135"} Oct 13 09:02:47 crc kubenswrapper[4685]: I1013 09:02:47.621107 4685 generic.go:334] "Generic (PLEG): container finished" podID="6a253ee4-12d7-4e32-9d53-27ecc16ce8cb" containerID="446f23b94fe81a8561329d0ad008a9ac7c533f1b4868061885fb9e65995cb656" exitCode=0 Oct 13 09:02:47 crc kubenswrapper[4685]: I1013 09:02:47.622764 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" event={"ID":"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb","Type":"ContainerDied","Data":"446f23b94fe81a8561329d0ad008a9ac7c533f1b4868061885fb9e65995cb656"} Oct 13 09:02:47 crc kubenswrapper[4685]: I1013 09:02:47.622798 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" event={"ID":"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb","Type":"ContainerStarted","Data":"f755cd6e41eba77fc76e07d259f1af977e371a9faa7beb1a67f52ca4ead82811"} Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.124843 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5746d5f994-pjpl9" podUID="9dd392f6-cf1c-4d69-81f8-167718c41926" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": read tcp 10.217.0.2:55614->10.217.0.163:9311: read: connection reset by peer" Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.125115 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5746d5f994-pjpl9" podUID="9dd392f6-cf1c-4d69-81f8-167718c41926" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": read tcp 10.217.0.2:55602->10.217.0.163:9311: read: connection reset by peer" Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.633515 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.638203 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" event={"ID":"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb","Type":"ContainerStarted","Data":"924dd8afb0dbcb29bcb97540614871285c070d93d517ed44e2aee8133da3bcfd"} Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.638394 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.643095 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"548142e8-94bd-4714-b65d-a4ec695000df","Type":"ContainerStarted","Data":"2ea1dee834b2cf10e4663b9d31c46ab5dadb18dcb165e4a28a827fffab9359fe"} Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.661231 4685 generic.go:334] "Generic (PLEG): container finished" podID="9dd392f6-cf1c-4d69-81f8-167718c41926" containerID="b559fba90dced4d180fb0896393fea41366c1de9c4acd2e31f5bf95bd3823d3b" exitCode=0 Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.661271 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5746d5f994-pjpl9" event={"ID":"9dd392f6-cf1c-4d69-81f8-167718c41926","Type":"ContainerDied","Data":"b559fba90dced4d180fb0896393fea41366c1de9c4acd2e31f5bf95bd3823d3b"} Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.661296 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5746d5f994-pjpl9" event={"ID":"9dd392f6-cf1c-4d69-81f8-167718c41926","Type":"ContainerDied","Data":"66b2160cade8f8daff442c987cd520da1b43ed981122997765c76f7f7d7a68ef"} Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.661314 4685 scope.go:117] "RemoveContainer" containerID="b559fba90dced4d180fb0896393fea41366c1de9c4acd2e31f5bf95bd3823d3b" Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.661407 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5746d5f994-pjpl9" Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.737525 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nkdq\" (UniqueName: \"kubernetes.io/projected/9dd392f6-cf1c-4d69-81f8-167718c41926-kube-api-access-8nkdq\") pod \"9dd392f6-cf1c-4d69-81f8-167718c41926\" (UID: \"9dd392f6-cf1c-4d69-81f8-167718c41926\") " Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.737577 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dd392f6-cf1c-4d69-81f8-167718c41926-config-data\") pod \"9dd392f6-cf1c-4d69-81f8-167718c41926\" (UID: \"9dd392f6-cf1c-4d69-81f8-167718c41926\") " Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.737687 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dd392f6-cf1c-4d69-81f8-167718c41926-logs\") pod \"9dd392f6-cf1c-4d69-81f8-167718c41926\" (UID: \"9dd392f6-cf1c-4d69-81f8-167718c41926\") " Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.737711 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9dd392f6-cf1c-4d69-81f8-167718c41926-config-data-custom\") pod \"9dd392f6-cf1c-4d69-81f8-167718c41926\" (UID: \"9dd392f6-cf1c-4d69-81f8-167718c41926\") " Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.737868 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dd392f6-cf1c-4d69-81f8-167718c41926-combined-ca-bundle\") pod \"9dd392f6-cf1c-4d69-81f8-167718c41926\" (UID: \"9dd392f6-cf1c-4d69-81f8-167718c41926\") " Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.744684 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9dd392f6-cf1c-4d69-81f8-167718c41926-logs" (OuterVolumeSpecName: "logs") pod "9dd392f6-cf1c-4d69-81f8-167718c41926" (UID: "9dd392f6-cf1c-4d69-81f8-167718c41926"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.759749 4685 scope.go:117] "RemoveContainer" containerID="cbe6d0a2722732a4d111f053aa7a593a39474ddfd32b4f71164834811c6471d1" Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.760943 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dd392f6-cf1c-4d69-81f8-167718c41926-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9dd392f6-cf1c-4d69-81f8-167718c41926" (UID: "9dd392f6-cf1c-4d69-81f8-167718c41926"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.761273 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dd392f6-cf1c-4d69-81f8-167718c41926-kube-api-access-8nkdq" (OuterVolumeSpecName: "kube-api-access-8nkdq") pod "9dd392f6-cf1c-4d69-81f8-167718c41926" (UID: "9dd392f6-cf1c-4d69-81f8-167718c41926"). InnerVolumeSpecName "kube-api-access-8nkdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.787232 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" podStartSLOduration=3.787215308 podStartE2EDuration="3.787215308s" podCreationTimestamp="2025-10-13 09:02:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:02:48.719291643 +0000 UTC m=+1093.867167404" watchObservedRunningTime="2025-10-13 09:02:48.787215308 +0000 UTC m=+1093.935091069" Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.791667 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.820594 4685 scope.go:117] "RemoveContainer" containerID="b559fba90dced4d180fb0896393fea41366c1de9c4acd2e31f5bf95bd3823d3b" Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.820782 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dd392f6-cf1c-4d69-81f8-167718c41926-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9dd392f6-cf1c-4d69-81f8-167718c41926" (UID: "9dd392f6-cf1c-4d69-81f8-167718c41926"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:48 crc kubenswrapper[4685]: E1013 09:02:48.821116 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b559fba90dced4d180fb0896393fea41366c1de9c4acd2e31f5bf95bd3823d3b\": container with ID starting with b559fba90dced4d180fb0896393fea41366c1de9c4acd2e31f5bf95bd3823d3b not found: ID does not exist" containerID="b559fba90dced4d180fb0896393fea41366c1de9c4acd2e31f5bf95bd3823d3b" Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.821156 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b559fba90dced4d180fb0896393fea41366c1de9c4acd2e31f5bf95bd3823d3b"} err="failed to get container status \"b559fba90dced4d180fb0896393fea41366c1de9c4acd2e31f5bf95bd3823d3b\": rpc error: code = NotFound desc = could not find container \"b559fba90dced4d180fb0896393fea41366c1de9c4acd2e31f5bf95bd3823d3b\": container with ID starting with b559fba90dced4d180fb0896393fea41366c1de9c4acd2e31f5bf95bd3823d3b not found: ID does not exist" Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.821181 4685 scope.go:117] "RemoveContainer" containerID="cbe6d0a2722732a4d111f053aa7a593a39474ddfd32b4f71164834811c6471d1" Oct 13 09:02:48 crc kubenswrapper[4685]: E1013 09:02:48.821584 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbe6d0a2722732a4d111f053aa7a593a39474ddfd32b4f71164834811c6471d1\": container with ID starting with cbe6d0a2722732a4d111f053aa7a593a39474ddfd32b4f71164834811c6471d1 not found: ID does not exist" containerID="cbe6d0a2722732a4d111f053aa7a593a39474ddfd32b4f71164834811c6471d1" Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.821607 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbe6d0a2722732a4d111f053aa7a593a39474ddfd32b4f71164834811c6471d1"} err="failed to get container status \"cbe6d0a2722732a4d111f053aa7a593a39474ddfd32b4f71164834811c6471d1\": rpc error: code = NotFound desc = could not find container \"cbe6d0a2722732a4d111f053aa7a593a39474ddfd32b4f71164834811c6471d1\": container with ID starting with cbe6d0a2722732a4d111f053aa7a593a39474ddfd32b4f71164834811c6471d1 not found: ID does not exist" Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.841124 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nkdq\" (UniqueName: \"kubernetes.io/projected/9dd392f6-cf1c-4d69-81f8-167718c41926-kube-api-access-8nkdq\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.841149 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dd392f6-cf1c-4d69-81f8-167718c41926-logs\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.841159 4685 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9dd392f6-cf1c-4d69-81f8-167718c41926-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.841167 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dd392f6-cf1c-4d69-81f8-167718c41926-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.849738 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dd392f6-cf1c-4d69-81f8-167718c41926-config-data" (OuterVolumeSpecName: "config-data") pod "9dd392f6-cf1c-4d69-81f8-167718c41926" (UID: "9dd392f6-cf1c-4d69-81f8-167718c41926"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:48 crc kubenswrapper[4685]: I1013 09:02:48.942478 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dd392f6-cf1c-4d69-81f8-167718c41926-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:49 crc kubenswrapper[4685]: I1013 09:02:49.106892 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5746d5f994-pjpl9"] Oct 13 09:02:49 crc kubenswrapper[4685]: I1013 09:02:49.117305 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5746d5f994-pjpl9"] Oct 13 09:02:49 crc kubenswrapper[4685]: I1013 09:02:49.532263 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dd392f6-cf1c-4d69-81f8-167718c41926" path="/var/lib/kubelet/pods/9dd392f6-cf1c-4d69-81f8-167718c41926/volumes" Oct 13 09:02:49 crc kubenswrapper[4685]: I1013 09:02:49.688559 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1bfd68fe-249f-43a2-8991-dbbe8eb5e184","Type":"ContainerStarted","Data":"f728afbb7b8b030fcb31c9b9ab7cfc037796a56edc52b6c9bbcca814498b1945"} Oct 13 09:02:49 crc kubenswrapper[4685]: I1013 09:02:49.759105 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-786f745fcf-5snx2" Oct 13 09:02:49 crc kubenswrapper[4685]: I1013 09:02:49.849100 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5c8fd5c86b-qkv7x"] Oct 13 09:02:49 crc kubenswrapper[4685]: I1013 09:02:49.854751 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5c8fd5c86b-qkv7x" podUID="89e018d5-d463-4b0e-8fef-f1d3230d703c" containerName="neutron-api" containerID="cri-o://39bc7e2bca44e0fe260292a0d1b6b6755d2e47c4bb3345e20a1f5eecbc4e2e70" gracePeriod=30 Oct 13 09:02:49 crc kubenswrapper[4685]: I1013 09:02:49.855299 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5c8fd5c86b-qkv7x" podUID="89e018d5-d463-4b0e-8fef-f1d3230d703c" containerName="neutron-httpd" containerID="cri-o://aec73af69652acd016a65ef6c7da86c1afc6374c84a7fc156440054ff0191bc3" gracePeriod=30 Oct 13 09:02:50 crc kubenswrapper[4685]: I1013 09:02:50.697662 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1bfd68fe-249f-43a2-8991-dbbe8eb5e184","Type":"ContainerStarted","Data":"b0fdac382829234a8d6f2ea2c079f694f05d5486eaa9f39bfaa6e9cdca83ece9"} Oct 13 09:02:50 crc kubenswrapper[4685]: I1013 09:02:50.699963 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"548142e8-94bd-4714-b65d-a4ec695000df","Type":"ContainerStarted","Data":"c4e5d25acb56a14f5e7c988629b99798caf9ed0571625e21721d9f721e1e717a"} Oct 13 09:02:50 crc kubenswrapper[4685]: I1013 09:02:50.700100 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="548142e8-94bd-4714-b65d-a4ec695000df" containerName="cinder-api-log" containerID="cri-o://2ea1dee834b2cf10e4663b9d31c46ab5dadb18dcb165e4a28a827fffab9359fe" gracePeriod=30 Oct 13 09:02:50 crc kubenswrapper[4685]: I1013 09:02:50.700189 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 13 09:02:50 crc kubenswrapper[4685]: I1013 09:02:50.700220 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="548142e8-94bd-4714-b65d-a4ec695000df" containerName="cinder-api" containerID="cri-o://c4e5d25acb56a14f5e7c988629b99798caf9ed0571625e21721d9f721e1e717a" gracePeriod=30 Oct 13 09:02:50 crc kubenswrapper[4685]: I1013 09:02:50.709694 4685 generic.go:334] "Generic (PLEG): container finished" podID="89e018d5-d463-4b0e-8fef-f1d3230d703c" containerID="aec73af69652acd016a65ef6c7da86c1afc6374c84a7fc156440054ff0191bc3" exitCode=0 Oct 13 09:02:50 crc kubenswrapper[4685]: I1013 09:02:50.709735 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c8fd5c86b-qkv7x" event={"ID":"89e018d5-d463-4b0e-8fef-f1d3230d703c","Type":"ContainerDied","Data":"aec73af69652acd016a65ef6c7da86c1afc6374c84a7fc156440054ff0191bc3"} Oct 13 09:02:50 crc kubenswrapper[4685]: I1013 09:02:50.726574 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.907350322 podStartE2EDuration="5.726556274s" podCreationTimestamp="2025-10-13 09:02:45 +0000 UTC" firstStartedPulling="2025-10-13 09:02:46.985565919 +0000 UTC m=+1092.133441680" lastFinishedPulling="2025-10-13 09:02:47.804771871 +0000 UTC m=+1092.952647632" observedRunningTime="2025-10-13 09:02:50.717428245 +0000 UTC m=+1095.865304006" watchObservedRunningTime="2025-10-13 09:02:50.726556274 +0000 UTC m=+1095.874432035" Oct 13 09:02:50 crc kubenswrapper[4685]: I1013 09:02:50.746494 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.746473447 podStartE2EDuration="4.746473447s" podCreationTimestamp="2025-10-13 09:02:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:02:50.736458053 +0000 UTC m=+1095.884333814" watchObservedRunningTime="2025-10-13 09:02:50.746473447 +0000 UTC m=+1095.894349208" Oct 13 09:02:51 crc kubenswrapper[4685]: I1013 09:02:51.225230 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 13 09:02:51 crc kubenswrapper[4685]: I1013 09:02:51.720299 4685 generic.go:334] "Generic (PLEG): container finished" podID="548142e8-94bd-4714-b65d-a4ec695000df" containerID="2ea1dee834b2cf10e4663b9d31c46ab5dadb18dcb165e4a28a827fffab9359fe" exitCode=143 Oct 13 09:02:51 crc kubenswrapper[4685]: I1013 09:02:51.720354 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"548142e8-94bd-4714-b65d-a4ec695000df","Type":"ContainerDied","Data":"2ea1dee834b2cf10e4663b9d31c46ab5dadb18dcb165e4a28a827fffab9359fe"} Oct 13 09:02:55 crc kubenswrapper[4685]: I1013 09:02:55.364329 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:56 crc kubenswrapper[4685]: I1013 09:02:56.400019 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:02:56 crc kubenswrapper[4685]: I1013 09:02:56.522251 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 13 09:02:56 crc kubenswrapper[4685]: I1013 09:02:56.542610 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-n8774"] Oct 13 09:02:56 crc kubenswrapper[4685]: I1013 09:02:56.543012 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-848cf88cfc-n8774" podUID="3688036a-1c46-45c2-ac9b-f2451a738f17" containerName="dnsmasq-dns" containerID="cri-o://79eed5a1d46f13a5aed18173e0129a27427bd7ef0b550f6470c04a7136b3abdd" gracePeriod=10 Oct 13 09:02:56 crc kubenswrapper[4685]: I1013 09:02:56.580047 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 13 09:02:56 crc kubenswrapper[4685]: I1013 09:02:56.839151 4685 generic.go:334] "Generic (PLEG): container finished" podID="3688036a-1c46-45c2-ac9b-f2451a738f17" containerID="79eed5a1d46f13a5aed18173e0129a27427bd7ef0b550f6470c04a7136b3abdd" exitCode=0 Oct 13 09:02:56 crc kubenswrapper[4685]: I1013 09:02:56.839689 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1bfd68fe-249f-43a2-8991-dbbe8eb5e184" containerName="cinder-scheduler" containerID="cri-o://f728afbb7b8b030fcb31c9b9ab7cfc037796a56edc52b6c9bbcca814498b1945" gracePeriod=30 Oct 13 09:02:56 crc kubenswrapper[4685]: I1013 09:02:56.840052 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-n8774" event={"ID":"3688036a-1c46-45c2-ac9b-f2451a738f17","Type":"ContainerDied","Data":"79eed5a1d46f13a5aed18173e0129a27427bd7ef0b550f6470c04a7136b3abdd"} Oct 13 09:02:56 crc kubenswrapper[4685]: I1013 09:02:56.840380 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1bfd68fe-249f-43a2-8991-dbbe8eb5e184" containerName="probe" containerID="cri-o://b0fdac382829234a8d6f2ea2c079f694f05d5486eaa9f39bfaa6e9cdca83ece9" gracePeriod=30 Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.235866 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.243186 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-ovsdbserver-nb\") pod \"3688036a-1c46-45c2-ac9b-f2451a738f17\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.243227 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-dns-swift-storage-0\") pod \"3688036a-1c46-45c2-ac9b-f2451a738f17\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.243301 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssdrm\" (UniqueName: \"kubernetes.io/projected/3688036a-1c46-45c2-ac9b-f2451a738f17-kube-api-access-ssdrm\") pod \"3688036a-1c46-45c2-ac9b-f2451a738f17\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.243334 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-config\") pod \"3688036a-1c46-45c2-ac9b-f2451a738f17\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.243405 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-ovsdbserver-sb\") pod \"3688036a-1c46-45c2-ac9b-f2451a738f17\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.243469 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-dns-svc\") pod \"3688036a-1c46-45c2-ac9b-f2451a738f17\" (UID: \"3688036a-1c46-45c2-ac9b-f2451a738f17\") " Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.263564 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3688036a-1c46-45c2-ac9b-f2451a738f17-kube-api-access-ssdrm" (OuterVolumeSpecName: "kube-api-access-ssdrm") pod "3688036a-1c46-45c2-ac9b-f2451a738f17" (UID: "3688036a-1c46-45c2-ac9b-f2451a738f17"). InnerVolumeSpecName "kube-api-access-ssdrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.326572 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3688036a-1c46-45c2-ac9b-f2451a738f17" (UID: "3688036a-1c46-45c2-ac9b-f2451a738f17"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.347698 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.347724 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssdrm\" (UniqueName: \"kubernetes.io/projected/3688036a-1c46-45c2-ac9b-f2451a738f17-kube-api-access-ssdrm\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.364425 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3688036a-1c46-45c2-ac9b-f2451a738f17" (UID: "3688036a-1c46-45c2-ac9b-f2451a738f17"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.374860 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3688036a-1c46-45c2-ac9b-f2451a738f17" (UID: "3688036a-1c46-45c2-ac9b-f2451a738f17"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.396105 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-config" (OuterVolumeSpecName: "config") pod "3688036a-1c46-45c2-ac9b-f2451a738f17" (UID: "3688036a-1c46-45c2-ac9b-f2451a738f17"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.408359 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3688036a-1c46-45c2-ac9b-f2451a738f17" (UID: "3688036a-1c46-45c2-ac9b-f2451a738f17"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.450640 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.450959 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.486422 4685 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.486906 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3688036a-1c46-45c2-ac9b-f2451a738f17-config\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.599544 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c8fd5c86b-qkv7x" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.623262 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-dbcf576b-msgft" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.696105 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-config\") pod \"89e018d5-d463-4b0e-8fef-f1d3230d703c\" (UID: \"89e018d5-d463-4b0e-8fef-f1d3230d703c\") " Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.696236 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-combined-ca-bundle\") pod \"89e018d5-d463-4b0e-8fef-f1d3230d703c\" (UID: \"89e018d5-d463-4b0e-8fef-f1d3230d703c\") " Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.696391 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-ovndb-tls-certs\") pod \"89e018d5-d463-4b0e-8fef-f1d3230d703c\" (UID: \"89e018d5-d463-4b0e-8fef-f1d3230d703c\") " Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.696490 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g426\" (UniqueName: \"kubernetes.io/projected/89e018d5-d463-4b0e-8fef-f1d3230d703c-kube-api-access-6g426\") pod \"89e018d5-d463-4b0e-8fef-f1d3230d703c\" (UID: \"89e018d5-d463-4b0e-8fef-f1d3230d703c\") " Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.696578 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-httpd-config\") pod \"89e018d5-d463-4b0e-8fef-f1d3230d703c\" (UID: \"89e018d5-d463-4b0e-8fef-f1d3230d703c\") " Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.737616 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89e018d5-d463-4b0e-8fef-f1d3230d703c-kube-api-access-6g426" (OuterVolumeSpecName: "kube-api-access-6g426") pod "89e018d5-d463-4b0e-8fef-f1d3230d703c" (UID: "89e018d5-d463-4b0e-8fef-f1d3230d703c"). InnerVolumeSpecName "kube-api-access-6g426". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.740421 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "89e018d5-d463-4b0e-8fef-f1d3230d703c" (UID: "89e018d5-d463-4b0e-8fef-f1d3230d703c"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.749374 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-db976b79d-q446d"] Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.749674 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-db976b79d-q446d" podUID="2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" containerName="horizon-log" containerID="cri-o://53e463eff1a6fb595b65a78ef21731ec4ee12f240e505e924644e92f36c5e9c4" gracePeriod=30 Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.749874 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-db976b79d-q446d" podUID="2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" containerName="horizon" containerID="cri-o://87642f946b3e4dd470d2686255f2a7452ddf07f367a7a524a315e59e389b70f5" gracePeriod=30 Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.799414 4685 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.799448 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g426\" (UniqueName: \"kubernetes.io/projected/89e018d5-d463-4b0e-8fef-f1d3230d703c-kube-api-access-6g426\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.881721 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-n8774" event={"ID":"3688036a-1c46-45c2-ac9b-f2451a738f17","Type":"ContainerDied","Data":"46a6aa94dbe5dd4aa8d17a3e63daccc62c2561e1b1865a4b7a0da44bccd1e07d"} Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.881848 4685 scope.go:117] "RemoveContainer" containerID="79eed5a1d46f13a5aed18173e0129a27427bd7ef0b550f6470c04a7136b3abdd" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.882036 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-n8774" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.904535 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-config" (OuterVolumeSpecName: "config") pod "89e018d5-d463-4b0e-8fef-f1d3230d703c" (UID: "89e018d5-d463-4b0e-8fef-f1d3230d703c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.930304 4685 generic.go:334] "Generic (PLEG): container finished" podID="89e018d5-d463-4b0e-8fef-f1d3230d703c" containerID="39bc7e2bca44e0fe260292a0d1b6b6755d2e47c4bb3345e20a1f5eecbc4e2e70" exitCode=0 Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.930446 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c8fd5c86b-qkv7x" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.930172 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "89e018d5-d463-4b0e-8fef-f1d3230d703c" (UID: "89e018d5-d463-4b0e-8fef-f1d3230d703c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.930743 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c8fd5c86b-qkv7x" event={"ID":"89e018d5-d463-4b0e-8fef-f1d3230d703c","Type":"ContainerDied","Data":"39bc7e2bca44e0fe260292a0d1b6b6755d2e47c4bb3345e20a1f5eecbc4e2e70"} Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.930832 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c8fd5c86b-qkv7x" event={"ID":"89e018d5-d463-4b0e-8fef-f1d3230d703c","Type":"ContainerDied","Data":"b929f4caf28b25b65a2d81186327ff1794a93ad7e305682a616a94112991ae7b"} Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.951342 4685 generic.go:334] "Generic (PLEG): container finished" podID="1bfd68fe-249f-43a2-8991-dbbe8eb5e184" containerID="b0fdac382829234a8d6f2ea2c079f694f05d5486eaa9f39bfaa6e9cdca83ece9" exitCode=0 Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.951373 4685 generic.go:334] "Generic (PLEG): container finished" podID="1bfd68fe-249f-43a2-8991-dbbe8eb5e184" containerID="f728afbb7b8b030fcb31c9b9ab7cfc037796a56edc52b6c9bbcca814498b1945" exitCode=0 Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.951398 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1bfd68fe-249f-43a2-8991-dbbe8eb5e184","Type":"ContainerDied","Data":"b0fdac382829234a8d6f2ea2c079f694f05d5486eaa9f39bfaa6e9cdca83ece9"} Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.951424 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1bfd68fe-249f-43a2-8991-dbbe8eb5e184","Type":"ContainerDied","Data":"f728afbb7b8b030fcb31c9b9ab7cfc037796a56edc52b6c9bbcca814498b1945"} Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.960098 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "89e018d5-d463-4b0e-8fef-f1d3230d703c" (UID: "89e018d5-d463-4b0e-8fef-f1d3230d703c"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.965005 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-n8774"] Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.969299 4685 scope.go:117] "RemoveContainer" containerID="49805c47194335609a4ffa0973190da27ed6a7b9578213b35d68b415372bef5d" Oct 13 09:02:57 crc kubenswrapper[4685]: I1013 09:02:57.971212 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-n8774"] Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.002030 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-config\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.002069 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.002079 4685 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/89e018d5-d463-4b0e-8fef-f1d3230d703c-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.017067 4685 scope.go:117] "RemoveContainer" containerID="aec73af69652acd016a65ef6c7da86c1afc6374c84a7fc156440054ff0191bc3" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.056684 4685 scope.go:117] "RemoveContainer" containerID="39bc7e2bca44e0fe260292a0d1b6b6755d2e47c4bb3345e20a1f5eecbc4e2e70" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.189541 4685 scope.go:117] "RemoveContainer" containerID="aec73af69652acd016a65ef6c7da86c1afc6374c84a7fc156440054ff0191bc3" Oct 13 09:02:58 crc kubenswrapper[4685]: E1013 09:02:58.190104 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aec73af69652acd016a65ef6c7da86c1afc6374c84a7fc156440054ff0191bc3\": container with ID starting with aec73af69652acd016a65ef6c7da86c1afc6374c84a7fc156440054ff0191bc3 not found: ID does not exist" containerID="aec73af69652acd016a65ef6c7da86c1afc6374c84a7fc156440054ff0191bc3" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.190146 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aec73af69652acd016a65ef6c7da86c1afc6374c84a7fc156440054ff0191bc3"} err="failed to get container status \"aec73af69652acd016a65ef6c7da86c1afc6374c84a7fc156440054ff0191bc3\": rpc error: code = NotFound desc = could not find container \"aec73af69652acd016a65ef6c7da86c1afc6374c84a7fc156440054ff0191bc3\": container with ID starting with aec73af69652acd016a65ef6c7da86c1afc6374c84a7fc156440054ff0191bc3 not found: ID does not exist" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.190167 4685 scope.go:117] "RemoveContainer" containerID="39bc7e2bca44e0fe260292a0d1b6b6755d2e47c4bb3345e20a1f5eecbc4e2e70" Oct 13 09:02:58 crc kubenswrapper[4685]: E1013 09:02:58.191352 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39bc7e2bca44e0fe260292a0d1b6b6755d2e47c4bb3345e20a1f5eecbc4e2e70\": container with ID starting with 39bc7e2bca44e0fe260292a0d1b6b6755d2e47c4bb3345e20a1f5eecbc4e2e70 not found: ID does not exist" containerID="39bc7e2bca44e0fe260292a0d1b6b6755d2e47c4bb3345e20a1f5eecbc4e2e70" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.191379 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39bc7e2bca44e0fe260292a0d1b6b6755d2e47c4bb3345e20a1f5eecbc4e2e70"} err="failed to get container status \"39bc7e2bca44e0fe260292a0d1b6b6755d2e47c4bb3345e20a1f5eecbc4e2e70\": rpc error: code = NotFound desc = could not find container \"39bc7e2bca44e0fe260292a0d1b6b6755d2e47c4bb3345e20a1f5eecbc4e2e70\": container with ID starting with 39bc7e2bca44e0fe260292a0d1b6b6755d2e47c4bb3345e20a1f5eecbc4e2e70 not found: ID does not exist" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.273437 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5c8fd5c86b-qkv7x"] Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.284948 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5c8fd5c86b-qkv7x"] Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.290743 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.424764 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-scripts\") pod \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.424879 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-config-data-custom\") pod \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.424925 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-combined-ca-bundle\") pod \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.425077 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vl84c\" (UniqueName: \"kubernetes.io/projected/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-kube-api-access-vl84c\") pod \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.425105 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-config-data\") pod \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.425141 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-etc-machine-id\") pod \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\" (UID: \"1bfd68fe-249f-43a2-8991-dbbe8eb5e184\") " Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.425505 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1bfd68fe-249f-43a2-8991-dbbe8eb5e184" (UID: "1bfd68fe-249f-43a2-8991-dbbe8eb5e184"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.429333 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1bfd68fe-249f-43a2-8991-dbbe8eb5e184" (UID: "1bfd68fe-249f-43a2-8991-dbbe8eb5e184"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.429627 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-kube-api-access-vl84c" (OuterVolumeSpecName: "kube-api-access-vl84c") pod "1bfd68fe-249f-43a2-8991-dbbe8eb5e184" (UID: "1bfd68fe-249f-43a2-8991-dbbe8eb5e184"). InnerVolumeSpecName "kube-api-access-vl84c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.430021 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-scripts" (OuterVolumeSpecName: "scripts") pod "1bfd68fe-249f-43a2-8991-dbbe8eb5e184" (UID: "1bfd68fe-249f-43a2-8991-dbbe8eb5e184"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.478275 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1bfd68fe-249f-43a2-8991-dbbe8eb5e184" (UID: "1bfd68fe-249f-43a2-8991-dbbe8eb5e184"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.527519 4685 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.527565 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.527579 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vl84c\" (UniqueName: \"kubernetes.io/projected/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-kube-api-access-vl84c\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.527593 4685 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.527606 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.540139 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-config-data" (OuterVolumeSpecName: "config-data") pod "1bfd68fe-249f-43a2-8991-dbbe8eb5e184" (UID: "1bfd68fe-249f-43a2-8991-dbbe8eb5e184"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.631469 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bfd68fe-249f-43a2-8991-dbbe8eb5e184-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.964028 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.964042 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1bfd68fe-249f-43a2-8991-dbbe8eb5e184","Type":"ContainerDied","Data":"e69d534659382dc4a7c4f1eb67bfcbf8e9de42febc83cf1a11a29c2068841ea9"} Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.964587 4685 scope.go:117] "RemoveContainer" containerID="b0fdac382829234a8d6f2ea2c079f694f05d5486eaa9f39bfaa6e9cdca83ece9" Oct 13 09:02:58 crc kubenswrapper[4685]: I1013 09:02:58.992692 4685 scope.go:117] "RemoveContainer" containerID="f728afbb7b8b030fcb31c9b9ab7cfc037796a56edc52b6c9bbcca814498b1945" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.002743 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.015031 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.038821 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 13 09:02:59 crc kubenswrapper[4685]: E1013 09:02:59.039281 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dd392f6-cf1c-4d69-81f8-167718c41926" containerName="barbican-api" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.039304 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dd392f6-cf1c-4d69-81f8-167718c41926" containerName="barbican-api" Oct 13 09:02:59 crc kubenswrapper[4685]: E1013 09:02:59.039326 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89e018d5-d463-4b0e-8fef-f1d3230d703c" containerName="neutron-api" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.039334 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="89e018d5-d463-4b0e-8fef-f1d3230d703c" containerName="neutron-api" Oct 13 09:02:59 crc kubenswrapper[4685]: E1013 09:02:59.039350 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89e018d5-d463-4b0e-8fef-f1d3230d703c" containerName="neutron-httpd" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.039357 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="89e018d5-d463-4b0e-8fef-f1d3230d703c" containerName="neutron-httpd" Oct 13 09:02:59 crc kubenswrapper[4685]: E1013 09:02:59.039378 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dd392f6-cf1c-4d69-81f8-167718c41926" containerName="barbican-api-log" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.039388 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dd392f6-cf1c-4d69-81f8-167718c41926" containerName="barbican-api-log" Oct 13 09:02:59 crc kubenswrapper[4685]: E1013 09:02:59.039404 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bfd68fe-249f-43a2-8991-dbbe8eb5e184" containerName="probe" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.039411 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bfd68fe-249f-43a2-8991-dbbe8eb5e184" containerName="probe" Oct 13 09:02:59 crc kubenswrapper[4685]: E1013 09:02:59.039428 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3688036a-1c46-45c2-ac9b-f2451a738f17" containerName="init" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.039435 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="3688036a-1c46-45c2-ac9b-f2451a738f17" containerName="init" Oct 13 09:02:59 crc kubenswrapper[4685]: E1013 09:02:59.039451 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bfd68fe-249f-43a2-8991-dbbe8eb5e184" containerName="cinder-scheduler" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.039458 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bfd68fe-249f-43a2-8991-dbbe8eb5e184" containerName="cinder-scheduler" Oct 13 09:02:59 crc kubenswrapper[4685]: E1013 09:02:59.039474 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3688036a-1c46-45c2-ac9b-f2451a738f17" containerName="dnsmasq-dns" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.039481 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="3688036a-1c46-45c2-ac9b-f2451a738f17" containerName="dnsmasq-dns" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.039701 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bfd68fe-249f-43a2-8991-dbbe8eb5e184" containerName="cinder-scheduler" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.039724 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bfd68fe-249f-43a2-8991-dbbe8eb5e184" containerName="probe" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.039739 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="89e018d5-d463-4b0e-8fef-f1d3230d703c" containerName="neutron-api" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.039758 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="3688036a-1c46-45c2-ac9b-f2451a738f17" containerName="dnsmasq-dns" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.039766 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="89e018d5-d463-4b0e-8fef-f1d3230d703c" containerName="neutron-httpd" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.039775 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dd392f6-cf1c-4d69-81f8-167718c41926" containerName="barbican-api-log" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.039792 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dd392f6-cf1c-4d69-81f8-167718c41926" containerName="barbican-api" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.040984 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.051258 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.065019 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.142093 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c60c26e9-8d95-47ec-9211-7053c42c471b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c60c26e9-8d95-47ec-9211-7053c42c471b\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.142354 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c60c26e9-8d95-47ec-9211-7053c42c471b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c60c26e9-8d95-47ec-9211-7053c42c471b\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.142459 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blbft\" (UniqueName: \"kubernetes.io/projected/c60c26e9-8d95-47ec-9211-7053c42c471b-kube-api-access-blbft\") pod \"cinder-scheduler-0\" (UID: \"c60c26e9-8d95-47ec-9211-7053c42c471b\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.142568 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c60c26e9-8d95-47ec-9211-7053c42c471b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c60c26e9-8d95-47ec-9211-7053c42c471b\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.142665 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c60c26e9-8d95-47ec-9211-7053c42c471b-scripts\") pod \"cinder-scheduler-0\" (UID: \"c60c26e9-8d95-47ec-9211-7053c42c471b\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.142766 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c60c26e9-8d95-47ec-9211-7053c42c471b-config-data\") pod \"cinder-scheduler-0\" (UID: \"c60c26e9-8d95-47ec-9211-7053c42c471b\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.265830 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c60c26e9-8d95-47ec-9211-7053c42c471b-scripts\") pod \"cinder-scheduler-0\" (UID: \"c60c26e9-8d95-47ec-9211-7053c42c471b\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.265929 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c60c26e9-8d95-47ec-9211-7053c42c471b-config-data\") pod \"cinder-scheduler-0\" (UID: \"c60c26e9-8d95-47ec-9211-7053c42c471b\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.266065 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c60c26e9-8d95-47ec-9211-7053c42c471b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c60c26e9-8d95-47ec-9211-7053c42c471b\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.266233 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c60c26e9-8d95-47ec-9211-7053c42c471b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c60c26e9-8d95-47ec-9211-7053c42c471b\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.266264 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blbft\" (UniqueName: \"kubernetes.io/projected/c60c26e9-8d95-47ec-9211-7053c42c471b-kube-api-access-blbft\") pod \"cinder-scheduler-0\" (UID: \"c60c26e9-8d95-47ec-9211-7053c42c471b\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.266324 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c60c26e9-8d95-47ec-9211-7053c42c471b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c60c26e9-8d95-47ec-9211-7053c42c471b\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.266437 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c60c26e9-8d95-47ec-9211-7053c42c471b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c60c26e9-8d95-47ec-9211-7053c42c471b\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.275227 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c60c26e9-8d95-47ec-9211-7053c42c471b-scripts\") pod \"cinder-scheduler-0\" (UID: \"c60c26e9-8d95-47ec-9211-7053c42c471b\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.275497 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c60c26e9-8d95-47ec-9211-7053c42c471b-config-data\") pod \"cinder-scheduler-0\" (UID: \"c60c26e9-8d95-47ec-9211-7053c42c471b\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.275748 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c60c26e9-8d95-47ec-9211-7053c42c471b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c60c26e9-8d95-47ec-9211-7053c42c471b\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.287075 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c60c26e9-8d95-47ec-9211-7053c42c471b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c60c26e9-8d95-47ec-9211-7053c42c471b\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.293663 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blbft\" (UniqueName: \"kubernetes.io/projected/c60c26e9-8d95-47ec-9211-7053c42c471b-kube-api-access-blbft\") pod \"cinder-scheduler-0\" (UID: \"c60c26e9-8d95-47ec-9211-7053c42c471b\") " pod="openstack/cinder-scheduler-0" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.374764 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.532350 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bfd68fe-249f-43a2-8991-dbbe8eb5e184" path="/var/lib/kubelet/pods/1bfd68fe-249f-43a2-8991-dbbe8eb5e184/volumes" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.533691 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3688036a-1c46-45c2-ac9b-f2451a738f17" path="/var/lib/kubelet/pods/3688036a-1c46-45c2-ac9b-f2451a738f17/volumes" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.534996 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89e018d5-d463-4b0e-8fef-f1d3230d703c" path="/var/lib/kubelet/pods/89e018d5-d463-4b0e-8fef-f1d3230d703c/volumes" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.734732 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.855963 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 13 09:02:59 crc kubenswrapper[4685]: I1013 09:02:59.975392 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c60c26e9-8d95-47ec-9211-7053c42c471b","Type":"ContainerStarted","Data":"53db866d1454120e597104c55317996cd9d3c52175dfaa8dc1cd49ac43710880"} Oct 13 09:03:00 crc kubenswrapper[4685]: I1013 09:03:00.987097 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c60c26e9-8d95-47ec-9211-7053c42c471b","Type":"ContainerStarted","Data":"ece626c5efb1dd395658e6f6d67a635993eb526cf67d91c7531d67efbd33213c"} Oct 13 09:03:01 crc kubenswrapper[4685]: I1013 09:03:01.028665 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-db976b79d-q446d" podUID="2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:49108->10.217.0.147:8443: read: connection reset by peer" Oct 13 09:03:01 crc kubenswrapper[4685]: I1013 09:03:01.997070 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c60c26e9-8d95-47ec-9211-7053c42c471b","Type":"ContainerStarted","Data":"98d4405ff1937d0fba93b4344bee70a217cdeb6a4d2c350732f2526acd9879bb"} Oct 13 09:03:01 crc kubenswrapper[4685]: I1013 09:03:01.999773 4685 generic.go:334] "Generic (PLEG): container finished" podID="2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" containerID="87642f946b3e4dd470d2686255f2a7452ddf07f367a7a524a315e59e389b70f5" exitCode=0 Oct 13 09:03:01 crc kubenswrapper[4685]: I1013 09:03:01.999819 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-db976b79d-q446d" event={"ID":"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc","Type":"ContainerDied","Data":"87642f946b3e4dd470d2686255f2a7452ddf07f367a7a524a315e59e389b70f5"} Oct 13 09:03:02 crc kubenswrapper[4685]: I1013 09:03:02.894897 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-79b7b9958d-7s7c7" Oct 13 09:03:02 crc kubenswrapper[4685]: I1013 09:03:02.912465 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.912448519 podStartE2EDuration="3.912448519s" podCreationTimestamp="2025-10-13 09:02:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:03:02.016352299 +0000 UTC m=+1107.164228060" watchObservedRunningTime="2025-10-13 09:03:02.912448519 +0000 UTC m=+1108.060324280" Oct 13 09:03:03 crc kubenswrapper[4685]: I1013 09:03:03.962226 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 13 09:03:03 crc kubenswrapper[4685]: I1013 09:03:03.963597 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 13 09:03:03 crc kubenswrapper[4685]: I1013 09:03:03.965282 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-5t962" Oct 13 09:03:03 crc kubenswrapper[4685]: I1013 09:03:03.965859 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 13 09:03:03 crc kubenswrapper[4685]: I1013 09:03:03.966208 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 13 09:03:03 crc kubenswrapper[4685]: I1013 09:03:03.973717 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 13 09:03:04 crc kubenswrapper[4685]: I1013 09:03:04.055607 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/31c9baa6-8272-45f1-b8d4-6c89f710fe20-openstack-config\") pod \"openstackclient\" (UID: \"31c9baa6-8272-45f1-b8d4-6c89f710fe20\") " pod="openstack/openstackclient" Oct 13 09:03:04 crc kubenswrapper[4685]: I1013 09:03:04.055718 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/31c9baa6-8272-45f1-b8d4-6c89f710fe20-openstack-config-secret\") pod \"openstackclient\" (UID: \"31c9baa6-8272-45f1-b8d4-6c89f710fe20\") " pod="openstack/openstackclient" Oct 13 09:03:04 crc kubenswrapper[4685]: I1013 09:03:04.055809 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31c9baa6-8272-45f1-b8d4-6c89f710fe20-combined-ca-bundle\") pod \"openstackclient\" (UID: \"31c9baa6-8272-45f1-b8d4-6c89f710fe20\") " pod="openstack/openstackclient" Oct 13 09:03:04 crc kubenswrapper[4685]: I1013 09:03:04.055827 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krqgd\" (UniqueName: \"kubernetes.io/projected/31c9baa6-8272-45f1-b8d4-6c89f710fe20-kube-api-access-krqgd\") pod \"openstackclient\" (UID: \"31c9baa6-8272-45f1-b8d4-6c89f710fe20\") " pod="openstack/openstackclient" Oct 13 09:03:04 crc kubenswrapper[4685]: I1013 09:03:04.157290 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31c9baa6-8272-45f1-b8d4-6c89f710fe20-combined-ca-bundle\") pod \"openstackclient\" (UID: \"31c9baa6-8272-45f1-b8d4-6c89f710fe20\") " pod="openstack/openstackclient" Oct 13 09:03:04 crc kubenswrapper[4685]: I1013 09:03:04.157336 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krqgd\" (UniqueName: \"kubernetes.io/projected/31c9baa6-8272-45f1-b8d4-6c89f710fe20-kube-api-access-krqgd\") pod \"openstackclient\" (UID: \"31c9baa6-8272-45f1-b8d4-6c89f710fe20\") " pod="openstack/openstackclient" Oct 13 09:03:04 crc kubenswrapper[4685]: I1013 09:03:04.157433 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/31c9baa6-8272-45f1-b8d4-6c89f710fe20-openstack-config\") pod \"openstackclient\" (UID: \"31c9baa6-8272-45f1-b8d4-6c89f710fe20\") " pod="openstack/openstackclient" Oct 13 09:03:04 crc kubenswrapper[4685]: I1013 09:03:04.157488 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/31c9baa6-8272-45f1-b8d4-6c89f710fe20-openstack-config-secret\") pod \"openstackclient\" (UID: \"31c9baa6-8272-45f1-b8d4-6c89f710fe20\") " pod="openstack/openstackclient" Oct 13 09:03:04 crc kubenswrapper[4685]: I1013 09:03:04.158429 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/31c9baa6-8272-45f1-b8d4-6c89f710fe20-openstack-config\") pod \"openstackclient\" (UID: \"31c9baa6-8272-45f1-b8d4-6c89f710fe20\") " pod="openstack/openstackclient" Oct 13 09:03:04 crc kubenswrapper[4685]: I1013 09:03:04.165480 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/31c9baa6-8272-45f1-b8d4-6c89f710fe20-openstack-config-secret\") pod \"openstackclient\" (UID: \"31c9baa6-8272-45f1-b8d4-6c89f710fe20\") " pod="openstack/openstackclient" Oct 13 09:03:04 crc kubenswrapper[4685]: I1013 09:03:04.165593 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31c9baa6-8272-45f1-b8d4-6c89f710fe20-combined-ca-bundle\") pod \"openstackclient\" (UID: \"31c9baa6-8272-45f1-b8d4-6c89f710fe20\") " pod="openstack/openstackclient" Oct 13 09:03:04 crc kubenswrapper[4685]: I1013 09:03:04.180625 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krqgd\" (UniqueName: \"kubernetes.io/projected/31c9baa6-8272-45f1-b8d4-6c89f710fe20-kube-api-access-krqgd\") pod \"openstackclient\" (UID: \"31c9baa6-8272-45f1-b8d4-6c89f710fe20\") " pod="openstack/openstackclient" Oct 13 09:03:04 crc kubenswrapper[4685]: I1013 09:03:04.285817 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 13 09:03:04 crc kubenswrapper[4685]: I1013 09:03:04.377011 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 13 09:03:04 crc kubenswrapper[4685]: I1013 09:03:04.763761 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 13 09:03:04 crc kubenswrapper[4685]: W1013 09:03:04.779865 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31c9baa6_8272_45f1_b8d4_6c89f710fe20.slice/crio-9f48b825a14e66060878cb0bac0a6ba22ea40fd17ecd6b114e76aaa0ff981fdf WatchSource:0}: Error finding container 9f48b825a14e66060878cb0bac0a6ba22ea40fd17ecd6b114e76aaa0ff981fdf: Status 404 returned error can't find the container with id 9f48b825a14e66060878cb0bac0a6ba22ea40fd17ecd6b114e76aaa0ff981fdf Oct 13 09:03:05 crc kubenswrapper[4685]: I1013 09:03:05.023443 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"31c9baa6-8272-45f1-b8d4-6c89f710fe20","Type":"ContainerStarted","Data":"9f48b825a14e66060878cb0bac0a6ba22ea40fd17ecd6b114e76aaa0ff981fdf"} Oct 13 09:03:08 crc kubenswrapper[4685]: I1013 09:03:08.139652 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 13 09:03:09 crc kubenswrapper[4685]: I1013 09:03:09.265654 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-db976b79d-q446d" podUID="2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 13 09:03:09 crc kubenswrapper[4685]: I1013 09:03:09.646862 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 13 09:03:11 crc kubenswrapper[4685]: I1013 09:03:11.977977 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-55864b6df5-xmtb6"] Oct 13 09:03:11 crc kubenswrapper[4685]: I1013 09:03:11.980882 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:11 crc kubenswrapper[4685]: I1013 09:03:11.984337 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 13 09:03:11 crc kubenswrapper[4685]: I1013 09:03:11.984606 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 13 09:03:11 crc kubenswrapper[4685]: I1013 09:03:11.991354 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.010709 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-55864b6df5-xmtb6"] Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.090342 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/86c3d3b4-cf99-46c5-b238-0efd9798f870-internal-tls-certs\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.090413 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86c3d3b4-cf99-46c5-b238-0efd9798f870-log-httpd\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.090474 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/86c3d3b4-cf99-46c5-b238-0efd9798f870-public-tls-certs\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.090500 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86c3d3b4-cf99-46c5-b238-0efd9798f870-combined-ca-bundle\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.090534 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86c3d3b4-cf99-46c5-b238-0efd9798f870-config-data\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.090581 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/86c3d3b4-cf99-46c5-b238-0efd9798f870-etc-swift\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.090600 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86c3d3b4-cf99-46c5-b238-0efd9798f870-run-httpd\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.090628 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxpq5\" (UniqueName: \"kubernetes.io/projected/86c3d3b4-cf99-46c5-b238-0efd9798f870-kube-api-access-rxpq5\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.192421 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86c3d3b4-cf99-46c5-b238-0efd9798f870-config-data\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.192498 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/86c3d3b4-cf99-46c5-b238-0efd9798f870-etc-swift\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.192519 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86c3d3b4-cf99-46c5-b238-0efd9798f870-run-httpd\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.192546 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxpq5\" (UniqueName: \"kubernetes.io/projected/86c3d3b4-cf99-46c5-b238-0efd9798f870-kube-api-access-rxpq5\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.192870 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/86c3d3b4-cf99-46c5-b238-0efd9798f870-internal-tls-certs\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.193105 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86c3d3b4-cf99-46c5-b238-0efd9798f870-run-httpd\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.193371 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86c3d3b4-cf99-46c5-b238-0efd9798f870-log-httpd\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.193390 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/86c3d3b4-cf99-46c5-b238-0efd9798f870-public-tls-certs\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.193414 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86c3d3b4-cf99-46c5-b238-0efd9798f870-combined-ca-bundle\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.196950 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86c3d3b4-cf99-46c5-b238-0efd9798f870-log-httpd\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.207135 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86c3d3b4-cf99-46c5-b238-0efd9798f870-config-data\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.212256 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/86c3d3b4-cf99-46c5-b238-0efd9798f870-etc-swift\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.219397 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/86c3d3b4-cf99-46c5-b238-0efd9798f870-public-tls-certs\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.219713 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxpq5\" (UniqueName: \"kubernetes.io/projected/86c3d3b4-cf99-46c5-b238-0efd9798f870-kube-api-access-rxpq5\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.236713 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/86c3d3b4-cf99-46c5-b238-0efd9798f870-internal-tls-certs\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.237354 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86c3d3b4-cf99-46c5-b238-0efd9798f870-combined-ca-bundle\") pod \"swift-proxy-55864b6df5-xmtb6\" (UID: \"86c3d3b4-cf99-46c5-b238-0efd9798f870\") " pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:12 crc kubenswrapper[4685]: I1013 09:03:12.315168 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:13 crc kubenswrapper[4685]: I1013 09:03:13.676460 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:13 crc kubenswrapper[4685]: I1013 09:03:13.676902 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7a6acbb5-144a-42d3-a770-60b0fe01de1f" containerName="ceilometer-central-agent" containerID="cri-o://d57b8e8b608b05b8d83161391dfa3b7f3461bc38b72ef48072d5ef4d5b6a8352" gracePeriod=30 Oct 13 09:03:13 crc kubenswrapper[4685]: I1013 09:03:13.677215 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7a6acbb5-144a-42d3-a770-60b0fe01de1f" containerName="sg-core" containerID="cri-o://da414155495a5e007b148dee7722c2f3ca121974df04632a00cae29c3944bfa4" gracePeriod=30 Oct 13 09:03:13 crc kubenswrapper[4685]: I1013 09:03:13.677226 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7a6acbb5-144a-42d3-a770-60b0fe01de1f" containerName="proxy-httpd" containerID="cri-o://ceb48dda53ea7f8d1d12ae899581138421aa4e83c663333e32936fb5902d73af" gracePeriod=30 Oct 13 09:03:13 crc kubenswrapper[4685]: I1013 09:03:13.677251 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7a6acbb5-144a-42d3-a770-60b0fe01de1f" containerName="ceilometer-notification-agent" containerID="cri-o://0824af0ad0d3226e908ed1d3fa419c6cb802a46b76ee0af084c4f6a3adfffe2f" gracePeriod=30 Oct 13 09:03:14 crc kubenswrapper[4685]: I1013 09:03:14.134892 4685 generic.go:334] "Generic (PLEG): container finished" podID="7a6acbb5-144a-42d3-a770-60b0fe01de1f" containerID="ceb48dda53ea7f8d1d12ae899581138421aa4e83c663333e32936fb5902d73af" exitCode=0 Oct 13 09:03:14 crc kubenswrapper[4685]: I1013 09:03:14.134940 4685 generic.go:334] "Generic (PLEG): container finished" podID="7a6acbb5-144a-42d3-a770-60b0fe01de1f" containerID="da414155495a5e007b148dee7722c2f3ca121974df04632a00cae29c3944bfa4" exitCode=2 Oct 13 09:03:14 crc kubenswrapper[4685]: I1013 09:03:14.134960 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a6acbb5-144a-42d3-a770-60b0fe01de1f","Type":"ContainerDied","Data":"ceb48dda53ea7f8d1d12ae899581138421aa4e83c663333e32936fb5902d73af"} Oct 13 09:03:14 crc kubenswrapper[4685]: I1013 09:03:14.134984 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a6acbb5-144a-42d3-a770-60b0fe01de1f","Type":"ContainerDied","Data":"da414155495a5e007b148dee7722c2f3ca121974df04632a00cae29c3944bfa4"} Oct 13 09:03:15 crc kubenswrapper[4685]: I1013 09:03:15.147287 4685 generic.go:334] "Generic (PLEG): container finished" podID="7a6acbb5-144a-42d3-a770-60b0fe01de1f" containerID="d57b8e8b608b05b8d83161391dfa3b7f3461bc38b72ef48072d5ef4d5b6a8352" exitCode=0 Oct 13 09:03:15 crc kubenswrapper[4685]: I1013 09:03:15.147330 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a6acbb5-144a-42d3-a770-60b0fe01de1f","Type":"ContainerDied","Data":"d57b8e8b608b05b8d83161391dfa3b7f3461bc38b72ef48072d5ef4d5b6a8352"} Oct 13 09:03:17 crc kubenswrapper[4685]: I1013 09:03:17.168664 4685 generic.go:334] "Generic (PLEG): container finished" podID="7a6acbb5-144a-42d3-a770-60b0fe01de1f" containerID="0824af0ad0d3226e908ed1d3fa419c6cb802a46b76ee0af084c4f6a3adfffe2f" exitCode=0 Oct 13 09:03:17 crc kubenswrapper[4685]: I1013 09:03:17.168733 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a6acbb5-144a-42d3-a770-60b0fe01de1f","Type":"ContainerDied","Data":"0824af0ad0d3226e908ed1d3fa419c6cb802a46b76ee0af084c4f6a3adfffe2f"} Oct 13 09:03:17 crc kubenswrapper[4685]: I1013 09:03:17.787841 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-sjt68"] Oct 13 09:03:17 crc kubenswrapper[4685]: I1013 09:03:17.789666 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sjt68" Oct 13 09:03:17 crc kubenswrapper[4685]: I1013 09:03:17.804468 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-sjt68"] Oct 13 09:03:17 crc kubenswrapper[4685]: I1013 09:03:17.882634 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-rtvbl"] Oct 13 09:03:17 crc kubenswrapper[4685]: I1013 09:03:17.883760 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-rtvbl" Oct 13 09:03:17 crc kubenswrapper[4685]: I1013 09:03:17.892532 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-rtvbl"] Oct 13 09:03:17 crc kubenswrapper[4685]: I1013 09:03:17.964982 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-kxkml"] Oct 13 09:03:17 crc kubenswrapper[4685]: I1013 09:03:17.965992 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kxkml" Oct 13 09:03:17 crc kubenswrapper[4685]: I1013 09:03:17.976649 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8rqt\" (UniqueName: \"kubernetes.io/projected/4119a02b-4484-438b-8521-33332cd63834-kube-api-access-p8rqt\") pod \"nova-api-db-create-sjt68\" (UID: \"4119a02b-4484-438b-8521-33332cd63834\") " pod="openstack/nova-api-db-create-sjt68" Oct 13 09:03:17 crc kubenswrapper[4685]: I1013 09:03:17.976705 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6jcl\" (UniqueName: \"kubernetes.io/projected/da42d77d-8d37-4e97-8852-2976d0be3012-kube-api-access-d6jcl\") pod \"nova-cell0-db-create-rtvbl\" (UID: \"da42d77d-8d37-4e97-8852-2976d0be3012\") " pod="openstack/nova-cell0-db-create-rtvbl" Oct 13 09:03:17 crc kubenswrapper[4685]: I1013 09:03:17.996328 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-kxkml"] Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.079432 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg7bx\" (UniqueName: \"kubernetes.io/projected/8826d2db-c766-469a-b840-83e177ebf61f-kube-api-access-lg7bx\") pod \"nova-cell1-db-create-kxkml\" (UID: \"8826d2db-c766-469a-b840-83e177ebf61f\") " pod="openstack/nova-cell1-db-create-kxkml" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.079517 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8rqt\" (UniqueName: \"kubernetes.io/projected/4119a02b-4484-438b-8521-33332cd63834-kube-api-access-p8rqt\") pod \"nova-api-db-create-sjt68\" (UID: \"4119a02b-4484-438b-8521-33332cd63834\") " pod="openstack/nova-api-db-create-sjt68" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.079576 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6jcl\" (UniqueName: \"kubernetes.io/projected/da42d77d-8d37-4e97-8852-2976d0be3012-kube-api-access-d6jcl\") pod \"nova-cell0-db-create-rtvbl\" (UID: \"da42d77d-8d37-4e97-8852-2976d0be3012\") " pod="openstack/nova-cell0-db-create-rtvbl" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.105391 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6jcl\" (UniqueName: \"kubernetes.io/projected/da42d77d-8d37-4e97-8852-2976d0be3012-kube-api-access-d6jcl\") pod \"nova-cell0-db-create-rtvbl\" (UID: \"da42d77d-8d37-4e97-8852-2976d0be3012\") " pod="openstack/nova-cell0-db-create-rtvbl" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.105847 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8rqt\" (UniqueName: \"kubernetes.io/projected/4119a02b-4484-438b-8521-33332cd63834-kube-api-access-p8rqt\") pod \"nova-api-db-create-sjt68\" (UID: \"4119a02b-4484-438b-8521-33332cd63834\") " pod="openstack/nova-api-db-create-sjt68" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.116896 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sjt68" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.193431 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg7bx\" (UniqueName: \"kubernetes.io/projected/8826d2db-c766-469a-b840-83e177ebf61f-kube-api-access-lg7bx\") pod \"nova-cell1-db-create-kxkml\" (UID: \"8826d2db-c766-469a-b840-83e177ebf61f\") " pod="openstack/nova-cell1-db-create-kxkml" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.205000 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-rtvbl" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.239315 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg7bx\" (UniqueName: \"kubernetes.io/projected/8826d2db-c766-469a-b840-83e177ebf61f-kube-api-access-lg7bx\") pod \"nova-cell1-db-create-kxkml\" (UID: \"8826d2db-c766-469a-b840-83e177ebf61f\") " pod="openstack/nova-cell1-db-create-kxkml" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.528894 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kxkml" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.535703 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.618486 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-scripts\") pod \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.618581 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-config-data\") pod \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.618637 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a6acbb5-144a-42d3-a770-60b0fe01de1f-run-httpd\") pod \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.618663 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwthr\" (UniqueName: \"kubernetes.io/projected/7a6acbb5-144a-42d3-a770-60b0fe01de1f-kube-api-access-bwthr\") pod \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.618683 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-sg-core-conf-yaml\") pod \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.618760 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a6acbb5-144a-42d3-a770-60b0fe01de1f-log-httpd\") pod \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.618861 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-combined-ca-bundle\") pod \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\" (UID: \"7a6acbb5-144a-42d3-a770-60b0fe01de1f\") " Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.619396 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a6acbb5-144a-42d3-a770-60b0fe01de1f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7a6acbb5-144a-42d3-a770-60b0fe01de1f" (UID: "7a6acbb5-144a-42d3-a770-60b0fe01de1f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.619654 4685 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a6acbb5-144a-42d3-a770-60b0fe01de1f-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.622296 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a6acbb5-144a-42d3-a770-60b0fe01de1f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7a6acbb5-144a-42d3-a770-60b0fe01de1f" (UID: "7a6acbb5-144a-42d3-a770-60b0fe01de1f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.626105 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a6acbb5-144a-42d3-a770-60b0fe01de1f-kube-api-access-bwthr" (OuterVolumeSpecName: "kube-api-access-bwthr") pod "7a6acbb5-144a-42d3-a770-60b0fe01de1f" (UID: "7a6acbb5-144a-42d3-a770-60b0fe01de1f"). InnerVolumeSpecName "kube-api-access-bwthr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.628742 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-scripts" (OuterVolumeSpecName: "scripts") pod "7a6acbb5-144a-42d3-a770-60b0fe01de1f" (UID: "7a6acbb5-144a-42d3-a770-60b0fe01de1f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.677058 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7a6acbb5-144a-42d3-a770-60b0fe01de1f" (UID: "7a6acbb5-144a-42d3-a770-60b0fe01de1f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.721782 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.721814 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwthr\" (UniqueName: \"kubernetes.io/projected/7a6acbb5-144a-42d3-a770-60b0fe01de1f-kube-api-access-bwthr\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.721828 4685 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.721839 4685 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a6acbb5-144a-42d3-a770-60b0fe01de1f-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.755036 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a6acbb5-144a-42d3-a770-60b0fe01de1f" (UID: "7a6acbb5-144a-42d3-a770-60b0fe01de1f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.780717 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-sjt68"] Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.823833 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.824079 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-config-data" (OuterVolumeSpecName: "config-data") pod "7a6acbb5-144a-42d3-a770-60b0fe01de1f" (UID: "7a6acbb5-144a-42d3-a770-60b0fe01de1f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.856325 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-55864b6df5-xmtb6"] Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.876601 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-rtvbl"] Oct 13 09:03:18 crc kubenswrapper[4685]: I1013 09:03:18.925895 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a6acbb5-144a-42d3-a770-60b0fe01de1f-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.122947 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-kxkml"] Oct 13 09:03:19 crc kubenswrapper[4685]: W1013 09:03:19.130798 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8826d2db_c766_469a_b840_83e177ebf61f.slice/crio-b9301cd2b8da38b638cd3bde312df932dc741ba92883381efbe58e99bac243c0 WatchSource:0}: Error finding container b9301cd2b8da38b638cd3bde312df932dc741ba92883381efbe58e99bac243c0: Status 404 returned error can't find the container with id b9301cd2b8da38b638cd3bde312df932dc741ba92883381efbe58e99bac243c0 Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.264461 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-db976b79d-q446d" podUID="2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.264782 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-db976b79d-q446d" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.277454 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a6acbb5-144a-42d3-a770-60b0fe01de1f","Type":"ContainerDied","Data":"3215d07ac65a0b9fbfc92d29b679c8b97f42551c1674456f66bea44f94feaa6a"} Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.277506 4685 scope.go:117] "RemoveContainer" containerID="ceb48dda53ea7f8d1d12ae899581138421aa4e83c663333e32936fb5902d73af" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.277637 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.286391 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"31c9baa6-8272-45f1-b8d4-6c89f710fe20","Type":"ContainerStarted","Data":"3b413ac4a8b184c6780c939e55df40379039ed8ce3d04feb62917f34b0081784"} Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.291440 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-55864b6df5-xmtb6" event={"ID":"86c3d3b4-cf99-46c5-b238-0efd9798f870","Type":"ContainerStarted","Data":"f91e3368a4921d4a6be60843a257e9c20d1534c77c028349efd7a2a473d9eaf1"} Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.297035 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kxkml" event={"ID":"8826d2db-c766-469a-b840-83e177ebf61f","Type":"ContainerStarted","Data":"b9301cd2b8da38b638cd3bde312df932dc741ba92883381efbe58e99bac243c0"} Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.298176 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-sjt68" event={"ID":"4119a02b-4484-438b-8521-33332cd63834","Type":"ContainerStarted","Data":"adcdab36d89fe8916d1019124684603c3aa616c3aa1f446de5b75ee4f6ae125c"} Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.301798 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-rtvbl" event={"ID":"da42d77d-8d37-4e97-8852-2976d0be3012","Type":"ContainerStarted","Data":"071d1cf163f3b317b21698ab59204fcb867e27d10d25f8b2c1aff8013eb3dba0"} Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.307031 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.9787581899999998 podStartE2EDuration="16.307009513s" podCreationTimestamp="2025-10-13 09:03:03 +0000 UTC" firstStartedPulling="2025-10-13 09:03:04.782930569 +0000 UTC m=+1109.930806330" lastFinishedPulling="2025-10-13 09:03:18.111181892 +0000 UTC m=+1123.259057653" observedRunningTime="2025-10-13 09:03:19.306624763 +0000 UTC m=+1124.454500524" watchObservedRunningTime="2025-10-13 09:03:19.307009513 +0000 UTC m=+1124.454885274" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.317430 4685 scope.go:117] "RemoveContainer" containerID="da414155495a5e007b148dee7722c2f3ca121974df04632a00cae29c3944bfa4" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.335947 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.347224 4685 scope.go:117] "RemoveContainer" containerID="0824af0ad0d3226e908ed1d3fa419c6cb802a46b76ee0af084c4f6a3adfffe2f" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.350803 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.360689 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:19 crc kubenswrapper[4685]: E1013 09:03:19.364530 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6acbb5-144a-42d3-a770-60b0fe01de1f" containerName="sg-core" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.364559 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6acbb5-144a-42d3-a770-60b0fe01de1f" containerName="sg-core" Oct 13 09:03:19 crc kubenswrapper[4685]: E1013 09:03:19.364571 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6acbb5-144a-42d3-a770-60b0fe01de1f" containerName="proxy-httpd" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.364579 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6acbb5-144a-42d3-a770-60b0fe01de1f" containerName="proxy-httpd" Oct 13 09:03:19 crc kubenswrapper[4685]: E1013 09:03:19.364595 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6acbb5-144a-42d3-a770-60b0fe01de1f" containerName="ceilometer-notification-agent" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.364601 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6acbb5-144a-42d3-a770-60b0fe01de1f" containerName="ceilometer-notification-agent" Oct 13 09:03:19 crc kubenswrapper[4685]: E1013 09:03:19.364616 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6acbb5-144a-42d3-a770-60b0fe01de1f" containerName="ceilometer-central-agent" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.364622 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6acbb5-144a-42d3-a770-60b0fe01de1f" containerName="ceilometer-central-agent" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.364840 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a6acbb5-144a-42d3-a770-60b0fe01de1f" containerName="ceilometer-central-agent" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.364857 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a6acbb5-144a-42d3-a770-60b0fe01de1f" containerName="ceilometer-notification-agent" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.364875 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a6acbb5-144a-42d3-a770-60b0fe01de1f" containerName="proxy-httpd" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.364888 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a6acbb5-144a-42d3-a770-60b0fe01de1f" containerName="sg-core" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.367666 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.377466 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.377508 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.391549 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.412727 4685 scope.go:117] "RemoveContainer" containerID="d57b8e8b608b05b8d83161391dfa3b7f3461bc38b72ef48072d5ef4d5b6a8352" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.435352 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.435423 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-log-httpd\") pod \"ceilometer-0\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.435450 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-scripts\") pod \"ceilometer-0\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.435480 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-config-data\") pod \"ceilometer-0\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.436030 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.436100 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhtps\" (UniqueName: \"kubernetes.io/projected/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-kube-api-access-qhtps\") pod \"ceilometer-0\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.436171 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-run-httpd\") pod \"ceilometer-0\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.519844 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a6acbb5-144a-42d3-a770-60b0fe01de1f" path="/var/lib/kubelet/pods/7a6acbb5-144a-42d3-a770-60b0fe01de1f/volumes" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.538572 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.538696 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-log-httpd\") pod \"ceilometer-0\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.538731 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-scripts\") pod \"ceilometer-0\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.538794 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-config-data\") pod \"ceilometer-0\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.538904 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.539093 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhtps\" (UniqueName: \"kubernetes.io/projected/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-kube-api-access-qhtps\") pod \"ceilometer-0\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.539152 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-run-httpd\") pod \"ceilometer-0\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.539882 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-run-httpd\") pod \"ceilometer-0\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.540713 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-log-httpd\") pod \"ceilometer-0\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.545285 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.552606 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-scripts\") pod \"ceilometer-0\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.553749 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.556087 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhtps\" (UniqueName: \"kubernetes.io/projected/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-kube-api-access-qhtps\") pod \"ceilometer-0\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.556966 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-config-data\") pod \"ceilometer-0\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " pod="openstack/ceilometer-0" Oct 13 09:03:19 crc kubenswrapper[4685]: I1013 09:03:19.689487 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:03:20 crc kubenswrapper[4685]: I1013 09:03:20.241021 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:20 crc kubenswrapper[4685]: I1013 09:03:20.314469 4685 generic.go:334] "Generic (PLEG): container finished" podID="4119a02b-4484-438b-8521-33332cd63834" containerID="03354a870611f8b31e4d0c070c0cf8a0b4ec8f54af85a611dbc62a161518f252" exitCode=0 Oct 13 09:03:20 crc kubenswrapper[4685]: I1013 09:03:20.314536 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-sjt68" event={"ID":"4119a02b-4484-438b-8521-33332cd63834","Type":"ContainerDied","Data":"03354a870611f8b31e4d0c070c0cf8a0b4ec8f54af85a611dbc62a161518f252"} Oct 13 09:03:20 crc kubenswrapper[4685]: I1013 09:03:20.316879 4685 generic.go:334] "Generic (PLEG): container finished" podID="da42d77d-8d37-4e97-8852-2976d0be3012" containerID="bcee7a4d5a46dd9d25c4ffcbe3993eecbcdb1a858588a54fcc9653da1624702b" exitCode=0 Oct 13 09:03:20 crc kubenswrapper[4685]: I1013 09:03:20.316953 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-rtvbl" event={"ID":"da42d77d-8d37-4e97-8852-2976d0be3012","Type":"ContainerDied","Data":"bcee7a4d5a46dd9d25c4ffcbe3993eecbcdb1a858588a54fcc9653da1624702b"} Oct 13 09:03:20 crc kubenswrapper[4685]: I1013 09:03:20.321654 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-55864b6df5-xmtb6" event={"ID":"86c3d3b4-cf99-46c5-b238-0efd9798f870","Type":"ContainerStarted","Data":"f8f968152c328c8a78462fbd575e1d4983f00086f373bc28beafa4f726e004d7"} Oct 13 09:03:20 crc kubenswrapper[4685]: I1013 09:03:20.321695 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-55864b6df5-xmtb6" event={"ID":"86c3d3b4-cf99-46c5-b238-0efd9798f870","Type":"ContainerStarted","Data":"39d28ebb224a8242c305a0d9610726806990d416e0b277b9630819b49a17c0ac"} Oct 13 09:03:20 crc kubenswrapper[4685]: I1013 09:03:20.322619 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:20 crc kubenswrapper[4685]: I1013 09:03:20.322662 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:20 crc kubenswrapper[4685]: I1013 09:03:20.325447 4685 generic.go:334] "Generic (PLEG): container finished" podID="8826d2db-c766-469a-b840-83e177ebf61f" containerID="a7eaea1b8683eb5def98f74abe4a0e16d74d1cc6b26849938ced531c5d1651a9" exitCode=0 Oct 13 09:03:20 crc kubenswrapper[4685]: I1013 09:03:20.325511 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kxkml" event={"ID":"8826d2db-c766-469a-b840-83e177ebf61f","Type":"ContainerDied","Data":"a7eaea1b8683eb5def98f74abe4a0e16d74d1cc6b26849938ced531c5d1651a9"} Oct 13 09:03:20 crc kubenswrapper[4685]: I1013 09:03:20.327831 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ea0f496-a1c1-41e9-9b72-d72a9acecb02","Type":"ContainerStarted","Data":"3dfc5b3352bf9e11e325ed55c8f2e00bb45712d784532d787a1e3109f41a0fc4"} Oct 13 09:03:20 crc kubenswrapper[4685]: I1013 09:03:20.374136 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-55864b6df5-xmtb6" podStartSLOduration=9.374119119 podStartE2EDuration="9.374119119s" podCreationTimestamp="2025-10-13 09:03:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:03:20.367066507 +0000 UTC m=+1125.514942268" watchObservedRunningTime="2025-10-13 09:03:20.374119119 +0000 UTC m=+1125.521994880" Oct 13 09:03:20 crc kubenswrapper[4685]: I1013 09:03:20.552153 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.353814 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ea0f496-a1c1-41e9-9b72-d72a9acecb02","Type":"ContainerStarted","Data":"014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825"} Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.363663 4685 generic.go:334] "Generic (PLEG): container finished" podID="548142e8-94bd-4714-b65d-a4ec695000df" containerID="c4e5d25acb56a14f5e7c988629b99798caf9ed0571625e21721d9f721e1e717a" exitCode=137 Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.363708 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"548142e8-94bd-4714-b65d-a4ec695000df","Type":"ContainerDied","Data":"c4e5d25acb56a14f5e7c988629b99798caf9ed0571625e21721d9f721e1e717a"} Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.363761 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"548142e8-94bd-4714-b65d-a4ec695000df","Type":"ContainerDied","Data":"c4f0a3d7ac361877440179c47d193b291be237e575dd0029c1520c931dd1a135"} Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.363773 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4f0a3d7ac361877440179c47d193b291be237e575dd0029c1520c931dd1a135" Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.419480 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.478669 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-scripts\") pod \"548142e8-94bd-4714-b65d-a4ec695000df\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.478729 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxpt4\" (UniqueName: \"kubernetes.io/projected/548142e8-94bd-4714-b65d-a4ec695000df-kube-api-access-bxpt4\") pod \"548142e8-94bd-4714-b65d-a4ec695000df\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.478792 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-combined-ca-bundle\") pod \"548142e8-94bd-4714-b65d-a4ec695000df\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.479943 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/548142e8-94bd-4714-b65d-a4ec695000df-logs\") pod \"548142e8-94bd-4714-b65d-a4ec695000df\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.479960 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/548142e8-94bd-4714-b65d-a4ec695000df-etc-machine-id\") pod \"548142e8-94bd-4714-b65d-a4ec695000df\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.479990 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-config-data\") pod \"548142e8-94bd-4714-b65d-a4ec695000df\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.480021 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-config-data-custom\") pod \"548142e8-94bd-4714-b65d-a4ec695000df\" (UID: \"548142e8-94bd-4714-b65d-a4ec695000df\") " Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.480940 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/548142e8-94bd-4714-b65d-a4ec695000df-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "548142e8-94bd-4714-b65d-a4ec695000df" (UID: "548142e8-94bd-4714-b65d-a4ec695000df"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.481207 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/548142e8-94bd-4714-b65d-a4ec695000df-logs" (OuterVolumeSpecName: "logs") pod "548142e8-94bd-4714-b65d-a4ec695000df" (UID: "548142e8-94bd-4714-b65d-a4ec695000df"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.499011 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "548142e8-94bd-4714-b65d-a4ec695000df" (UID: "548142e8-94bd-4714-b65d-a4ec695000df"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.506055 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-scripts" (OuterVolumeSpecName: "scripts") pod "548142e8-94bd-4714-b65d-a4ec695000df" (UID: "548142e8-94bd-4714-b65d-a4ec695000df"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.521852 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/548142e8-94bd-4714-b65d-a4ec695000df-kube-api-access-bxpt4" (OuterVolumeSpecName: "kube-api-access-bxpt4") pod "548142e8-94bd-4714-b65d-a4ec695000df" (UID: "548142e8-94bd-4714-b65d-a4ec695000df"). InnerVolumeSpecName "kube-api-access-bxpt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.560275 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "548142e8-94bd-4714-b65d-a4ec695000df" (UID: "548142e8-94bd-4714-b65d-a4ec695000df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.582161 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-config-data" (OuterVolumeSpecName: "config-data") pod "548142e8-94bd-4714-b65d-a4ec695000df" (UID: "548142e8-94bd-4714-b65d-a4ec695000df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.582682 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.582787 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxpt4\" (UniqueName: \"kubernetes.io/projected/548142e8-94bd-4714-b65d-a4ec695000df-kube-api-access-bxpt4\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.582896 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.582992 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/548142e8-94bd-4714-b65d-a4ec695000df-logs\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.591595 4685 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/548142e8-94bd-4714-b65d-a4ec695000df-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.591691 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.591748 4685 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/548142e8-94bd-4714-b65d-a4ec695000df-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.716009 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kxkml" Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.801869 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lg7bx\" (UniqueName: \"kubernetes.io/projected/8826d2db-c766-469a-b840-83e177ebf61f-kube-api-access-lg7bx\") pod \"8826d2db-c766-469a-b840-83e177ebf61f\" (UID: \"8826d2db-c766-469a-b840-83e177ebf61f\") " Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.806716 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8826d2db-c766-469a-b840-83e177ebf61f-kube-api-access-lg7bx" (OuterVolumeSpecName: "kube-api-access-lg7bx") pod "8826d2db-c766-469a-b840-83e177ebf61f" (UID: "8826d2db-c766-469a-b840-83e177ebf61f"). InnerVolumeSpecName "kube-api-access-lg7bx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.907225 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lg7bx\" (UniqueName: \"kubernetes.io/projected/8826d2db-c766-469a-b840-83e177ebf61f-kube-api-access-lg7bx\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:21 crc kubenswrapper[4685]: I1013 09:03:21.959089 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sjt68" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.111770 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8rqt\" (UniqueName: \"kubernetes.io/projected/4119a02b-4484-438b-8521-33332cd63834-kube-api-access-p8rqt\") pod \"4119a02b-4484-438b-8521-33332cd63834\" (UID: \"4119a02b-4484-438b-8521-33332cd63834\") " Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.127366 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4119a02b-4484-438b-8521-33332cd63834-kube-api-access-p8rqt" (OuterVolumeSpecName: "kube-api-access-p8rqt") pod "4119a02b-4484-438b-8521-33332cd63834" (UID: "4119a02b-4484-438b-8521-33332cd63834"). InnerVolumeSpecName "kube-api-access-p8rqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.213566 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8rqt\" (UniqueName: \"kubernetes.io/projected/4119a02b-4484-438b-8521-33332cd63834-kube-api-access-p8rqt\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.241630 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-rtvbl" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.371899 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kxkml" event={"ID":"8826d2db-c766-469a-b840-83e177ebf61f","Type":"ContainerDied","Data":"b9301cd2b8da38b638cd3bde312df932dc741ba92883381efbe58e99bac243c0"} Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.372175 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9301cd2b8da38b638cd3bde312df932dc741ba92883381efbe58e99bac243c0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.372221 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kxkml" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.374629 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-sjt68" event={"ID":"4119a02b-4484-438b-8521-33332cd63834","Type":"ContainerDied","Data":"adcdab36d89fe8916d1019124684603c3aa616c3aa1f446de5b75ee4f6ae125c"} Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.374654 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="adcdab36d89fe8916d1019124684603c3aa616c3aa1f446de5b75ee4f6ae125c" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.374695 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sjt68" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.382578 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.382625 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-rtvbl" event={"ID":"da42d77d-8d37-4e97-8852-2976d0be3012","Type":"ContainerDied","Data":"071d1cf163f3b317b21698ab59204fcb867e27d10d25f8b2c1aff8013eb3dba0"} Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.382676 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="071d1cf163f3b317b21698ab59204fcb867e27d10d25f8b2c1aff8013eb3dba0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.383110 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-rtvbl" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.415773 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6jcl\" (UniqueName: \"kubernetes.io/projected/da42d77d-8d37-4e97-8852-2976d0be3012-kube-api-access-d6jcl\") pod \"da42d77d-8d37-4e97-8852-2976d0be3012\" (UID: \"da42d77d-8d37-4e97-8852-2976d0be3012\") " Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.428740 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.436219 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.439882 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da42d77d-8d37-4e97-8852-2976d0be3012-kube-api-access-d6jcl" (OuterVolumeSpecName: "kube-api-access-d6jcl") pod "da42d77d-8d37-4e97-8852-2976d0be3012" (UID: "da42d77d-8d37-4e97-8852-2976d0be3012"). InnerVolumeSpecName "kube-api-access-d6jcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.465451 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 13 09:03:22 crc kubenswrapper[4685]: E1013 09:03:22.465950 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="548142e8-94bd-4714-b65d-a4ec695000df" containerName="cinder-api-log" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.465969 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="548142e8-94bd-4714-b65d-a4ec695000df" containerName="cinder-api-log" Oct 13 09:03:22 crc kubenswrapper[4685]: E1013 09:03:22.465991 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="548142e8-94bd-4714-b65d-a4ec695000df" containerName="cinder-api" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.466014 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="548142e8-94bd-4714-b65d-a4ec695000df" containerName="cinder-api" Oct 13 09:03:22 crc kubenswrapper[4685]: E1013 09:03:22.466027 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4119a02b-4484-438b-8521-33332cd63834" containerName="mariadb-database-create" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.466034 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="4119a02b-4484-438b-8521-33332cd63834" containerName="mariadb-database-create" Oct 13 09:03:22 crc kubenswrapper[4685]: E1013 09:03:22.466043 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da42d77d-8d37-4e97-8852-2976d0be3012" containerName="mariadb-database-create" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.466050 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="da42d77d-8d37-4e97-8852-2976d0be3012" containerName="mariadb-database-create" Oct 13 09:03:22 crc kubenswrapper[4685]: E1013 09:03:22.466270 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8826d2db-c766-469a-b840-83e177ebf61f" containerName="mariadb-database-create" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.466279 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="8826d2db-c766-469a-b840-83e177ebf61f" containerName="mariadb-database-create" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.466517 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="8826d2db-c766-469a-b840-83e177ebf61f" containerName="mariadb-database-create" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.470488 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="548142e8-94bd-4714-b65d-a4ec695000df" containerName="cinder-api" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.470521 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="4119a02b-4484-438b-8521-33332cd63834" containerName="mariadb-database-create" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.470537 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="548142e8-94bd-4714-b65d-a4ec695000df" containerName="cinder-api-log" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.470551 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="da42d77d-8d37-4e97-8852-2976d0be3012" containerName="mariadb-database-create" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.471700 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.476603 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.476817 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.476827 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.479174 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.518145 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6jcl\" (UniqueName: \"kubernetes.io/projected/da42d77d-8d37-4e97-8852-2976d0be3012-kube-api-access-d6jcl\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.619260 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8z5x\" (UniqueName: \"kubernetes.io/projected/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-kube-api-access-w8z5x\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.619366 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.619399 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-config-data-custom\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.619423 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-scripts\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.619503 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-logs\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.619548 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.619565 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.619597 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.619636 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-config-data\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.721297 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.721342 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-config-data-custom\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.721362 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-scripts\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.721422 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-logs\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.721459 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.721476 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.721513 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.721558 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-config-data\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.721561 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.721635 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8z5x\" (UniqueName: \"kubernetes.io/projected/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-kube-api-access-w8z5x\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.722349 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-logs\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.725967 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-scripts\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.726590 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-config-data-custom\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.726601 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.727004 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.727777 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.732066 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-config-data\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.740073 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8z5x\" (UniqueName: \"kubernetes.io/projected/9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f-kube-api-access-w8z5x\") pod \"cinder-api-0\" (UID: \"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f\") " pod="openstack/cinder-api-0" Oct 13 09:03:22 crc kubenswrapper[4685]: I1013 09:03:22.795192 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 13 09:03:23 crc kubenswrapper[4685]: I1013 09:03:23.392421 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ea0f496-a1c1-41e9-9b72-d72a9acecb02","Type":"ContainerStarted","Data":"67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8"} Oct 13 09:03:23 crc kubenswrapper[4685]: I1013 09:03:23.470643 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 13 09:03:23 crc kubenswrapper[4685]: I1013 09:03:23.541725 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="548142e8-94bd-4714-b65d-a4ec695000df" path="/var/lib/kubelet/pods/548142e8-94bd-4714-b65d-a4ec695000df/volumes" Oct 13 09:03:24 crc kubenswrapper[4685]: I1013 09:03:24.417156 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ea0f496-a1c1-41e9-9b72-d72a9acecb02","Type":"ContainerStarted","Data":"5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4"} Oct 13 09:03:24 crc kubenswrapper[4685]: I1013 09:03:24.421212 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f","Type":"ContainerStarted","Data":"4618238174aeaec42c783843ad917b233cf39682ccb631080369e71464b37e9f"} Oct 13 09:03:24 crc kubenswrapper[4685]: I1013 09:03:24.421257 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f","Type":"ContainerStarted","Data":"e1b6e75eb0ec1e31e1f88f39e6d862811987f5d1c19742e0f2e243ba463ee486"} Oct 13 09:03:25 crc kubenswrapper[4685]: I1013 09:03:25.111712 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 09:03:25 crc kubenswrapper[4685]: I1013 09:03:25.112579 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="babeefc6-9cbe-4f6c-9487-731a9270c1c4" containerName="glance-log" containerID="cri-o://1151b5de139a4c03b47ac9c2ee15276efe6b1481c2f0cba8e9123177807e5804" gracePeriod=30 Oct 13 09:03:25 crc kubenswrapper[4685]: I1013 09:03:25.112850 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="babeefc6-9cbe-4f6c-9487-731a9270c1c4" containerName="glance-httpd" containerID="cri-o://3346630c966f3d7aff2a38529b3649c2ce0e925ee871b7630194486f5d391a51" gracePeriod=30 Oct 13 09:03:25 crc kubenswrapper[4685]: I1013 09:03:25.442100 4685 generic.go:334] "Generic (PLEG): container finished" podID="babeefc6-9cbe-4f6c-9487-731a9270c1c4" containerID="1151b5de139a4c03b47ac9c2ee15276efe6b1481c2f0cba8e9123177807e5804" exitCode=143 Oct 13 09:03:25 crc kubenswrapper[4685]: I1013 09:03:25.442216 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"babeefc6-9cbe-4f6c-9487-731a9270c1c4","Type":"ContainerDied","Data":"1151b5de139a4c03b47ac9c2ee15276efe6b1481c2f0cba8e9123177807e5804"} Oct 13 09:03:25 crc kubenswrapper[4685]: I1013 09:03:25.449235 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f","Type":"ContainerStarted","Data":"4a8fb73c54d8235627adaf561c4bab5f8ff554d891727e42277377e4f02369f2"} Oct 13 09:03:25 crc kubenswrapper[4685]: I1013 09:03:25.449420 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 13 09:03:25 crc kubenswrapper[4685]: I1013 09:03:25.470336 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.470326423 podStartE2EDuration="3.470326423s" podCreationTimestamp="2025-10-13 09:03:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:03:25.468435512 +0000 UTC m=+1130.616311273" watchObservedRunningTime="2025-10-13 09:03:25.470326423 +0000 UTC m=+1130.618202184" Oct 13 09:03:26 crc kubenswrapper[4685]: I1013 09:03:26.472560 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ea0f496-a1c1-41e9-9b72-d72a9acecb02","Type":"ContainerStarted","Data":"667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2"} Oct 13 09:03:26 crc kubenswrapper[4685]: I1013 09:03:26.472870 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1ea0f496-a1c1-41e9-9b72-d72a9acecb02" containerName="ceilometer-central-agent" containerID="cri-o://014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825" gracePeriod=30 Oct 13 09:03:26 crc kubenswrapper[4685]: I1013 09:03:26.472949 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 13 09:03:26 crc kubenswrapper[4685]: I1013 09:03:26.473051 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1ea0f496-a1c1-41e9-9b72-d72a9acecb02" containerName="proxy-httpd" containerID="cri-o://667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2" gracePeriod=30 Oct 13 09:03:26 crc kubenswrapper[4685]: I1013 09:03:26.473128 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1ea0f496-a1c1-41e9-9b72-d72a9acecb02" containerName="sg-core" containerID="cri-o://5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4" gracePeriod=30 Oct 13 09:03:26 crc kubenswrapper[4685]: I1013 09:03:26.473176 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1ea0f496-a1c1-41e9-9b72-d72a9acecb02" containerName="ceilometer-notification-agent" containerID="cri-o://67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8" gracePeriod=30 Oct 13 09:03:26 crc kubenswrapper[4685]: I1013 09:03:26.514736 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.415426763 podStartE2EDuration="7.51471081s" podCreationTimestamp="2025-10-13 09:03:19 +0000 UTC" firstStartedPulling="2025-10-13 09:03:20.273121919 +0000 UTC m=+1125.420997680" lastFinishedPulling="2025-10-13 09:03:25.372405966 +0000 UTC m=+1130.520281727" observedRunningTime="2025-10-13 09:03:26.499464366 +0000 UTC m=+1131.647340127" watchObservedRunningTime="2025-10-13 09:03:26.51471081 +0000 UTC m=+1131.662586591" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.252291 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.307154 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-combined-ca-bundle\") pod \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.307209 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-config-data\") pod \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.307233 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-sg-core-conf-yaml\") pod \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.307279 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-log-httpd\") pod \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.307395 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhtps\" (UniqueName: \"kubernetes.io/projected/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-kube-api-access-qhtps\") pod \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.307420 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-scripts\") pod \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.307467 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-run-httpd\") pod \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\" (UID: \"1ea0f496-a1c1-41e9-9b72-d72a9acecb02\") " Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.308232 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1ea0f496-a1c1-41e9-9b72-d72a9acecb02" (UID: "1ea0f496-a1c1-41e9-9b72-d72a9acecb02"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.309414 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1ea0f496-a1c1-41e9-9b72-d72a9acecb02" (UID: "1ea0f496-a1c1-41e9-9b72-d72a9acecb02"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.316906 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-kube-api-access-qhtps" (OuterVolumeSpecName: "kube-api-access-qhtps") pod "1ea0f496-a1c1-41e9-9b72-d72a9acecb02" (UID: "1ea0f496-a1c1-41e9-9b72-d72a9acecb02"). InnerVolumeSpecName "kube-api-access-qhtps". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.317906 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-scripts" (OuterVolumeSpecName: "scripts") pod "1ea0f496-a1c1-41e9-9b72-d72a9acecb02" (UID: "1ea0f496-a1c1-41e9-9b72-d72a9acecb02"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.329472 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.333805 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-55864b6df5-xmtb6" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.359190 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1ea0f496-a1c1-41e9-9b72-d72a9acecb02" (UID: "1ea0f496-a1c1-41e9-9b72-d72a9acecb02"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.409645 4685 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.409672 4685 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.409681 4685 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.409689 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhtps\" (UniqueName: \"kubernetes.io/projected/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-kube-api-access-qhtps\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.409697 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.443566 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ea0f496-a1c1-41e9-9b72-d72a9acecb02" (UID: "1ea0f496-a1c1-41e9-9b72-d72a9acecb02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.482177 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-config-data" (OuterVolumeSpecName: "config-data") pod "1ea0f496-a1c1-41e9-9b72-d72a9acecb02" (UID: "1ea0f496-a1c1-41e9-9b72-d72a9acecb02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.488347 4685 generic.go:334] "Generic (PLEG): container finished" podID="1ea0f496-a1c1-41e9-9b72-d72a9acecb02" containerID="667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2" exitCode=0 Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.488380 4685 generic.go:334] "Generic (PLEG): container finished" podID="1ea0f496-a1c1-41e9-9b72-d72a9acecb02" containerID="5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4" exitCode=2 Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.488387 4685 generic.go:334] "Generic (PLEG): container finished" podID="1ea0f496-a1c1-41e9-9b72-d72a9acecb02" containerID="67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8" exitCode=0 Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.488394 4685 generic.go:334] "Generic (PLEG): container finished" podID="1ea0f496-a1c1-41e9-9b72-d72a9acecb02" containerID="014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825" exitCode=0 Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.489148 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.493079 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ea0f496-a1c1-41e9-9b72-d72a9acecb02","Type":"ContainerDied","Data":"667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2"} Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.493150 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ea0f496-a1c1-41e9-9b72-d72a9acecb02","Type":"ContainerDied","Data":"5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4"} Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.493163 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ea0f496-a1c1-41e9-9b72-d72a9acecb02","Type":"ContainerDied","Data":"67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8"} Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.493173 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ea0f496-a1c1-41e9-9b72-d72a9acecb02","Type":"ContainerDied","Data":"014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825"} Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.493185 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ea0f496-a1c1-41e9-9b72-d72a9acecb02","Type":"ContainerDied","Data":"3dfc5b3352bf9e11e325ed55c8f2e00bb45712d784532d787a1e3109f41a0fc4"} Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.493203 4685 scope.go:117] "RemoveContainer" containerID="667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.517478 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.518138 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ea0f496-a1c1-41e9-9b72-d72a9acecb02-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.544150 4685 scope.go:117] "RemoveContainer" containerID="5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.568466 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.586849 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.597949 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:27 crc kubenswrapper[4685]: E1013 09:03:27.598302 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea0f496-a1c1-41e9-9b72-d72a9acecb02" containerName="sg-core" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.598319 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea0f496-a1c1-41e9-9b72-d72a9acecb02" containerName="sg-core" Oct 13 09:03:27 crc kubenswrapper[4685]: E1013 09:03:27.598327 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea0f496-a1c1-41e9-9b72-d72a9acecb02" containerName="ceilometer-central-agent" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.598333 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea0f496-a1c1-41e9-9b72-d72a9acecb02" containerName="ceilometer-central-agent" Oct 13 09:03:27 crc kubenswrapper[4685]: E1013 09:03:27.598350 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea0f496-a1c1-41e9-9b72-d72a9acecb02" containerName="proxy-httpd" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.598357 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea0f496-a1c1-41e9-9b72-d72a9acecb02" containerName="proxy-httpd" Oct 13 09:03:27 crc kubenswrapper[4685]: E1013 09:03:27.598369 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea0f496-a1c1-41e9-9b72-d72a9acecb02" containerName="ceilometer-notification-agent" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.598374 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea0f496-a1c1-41e9-9b72-d72a9acecb02" containerName="ceilometer-notification-agent" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.598527 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ea0f496-a1c1-41e9-9b72-d72a9acecb02" containerName="ceilometer-notification-agent" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.598552 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ea0f496-a1c1-41e9-9b72-d72a9acecb02" containerName="ceilometer-central-agent" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.598566 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ea0f496-a1c1-41e9-9b72-d72a9acecb02" containerName="proxy-httpd" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.598579 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ea0f496-a1c1-41e9-9b72-d72a9acecb02" containerName="sg-core" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.600058 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.607406 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.607621 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.607867 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.662491 4685 scope.go:117] "RemoveContainer" containerID="67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.689833 4685 scope.go:117] "RemoveContainer" containerID="014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.714145 4685 scope.go:117] "RemoveContainer" containerID="667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2" Oct 13 09:03:27 crc kubenswrapper[4685]: E1013 09:03:27.714575 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2\": container with ID starting with 667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2 not found: ID does not exist" containerID="667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.714622 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2"} err="failed to get container status \"667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2\": rpc error: code = NotFound desc = could not find container \"667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2\": container with ID starting with 667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2 not found: ID does not exist" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.714643 4685 scope.go:117] "RemoveContainer" containerID="5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4" Oct 13 09:03:27 crc kubenswrapper[4685]: E1013 09:03:27.718002 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4\": container with ID starting with 5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4 not found: ID does not exist" containerID="5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.718043 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4"} err="failed to get container status \"5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4\": rpc error: code = NotFound desc = could not find container \"5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4\": container with ID starting with 5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4 not found: ID does not exist" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.718058 4685 scope.go:117] "RemoveContainer" containerID="67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8" Oct 13 09:03:27 crc kubenswrapper[4685]: E1013 09:03:27.721753 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8\": container with ID starting with 67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8 not found: ID does not exist" containerID="67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.721875 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8"} err="failed to get container status \"67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8\": rpc error: code = NotFound desc = could not find container \"67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8\": container with ID starting with 67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8 not found: ID does not exist" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.721897 4685 scope.go:117] "RemoveContainer" containerID="014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.721957 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-scripts\") pod \"ceilometer-0\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.722087 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pvpb\" (UniqueName: \"kubernetes.io/projected/385f189a-9a2f-42db-a4c1-dd7b332a4f91-kube-api-access-5pvpb\") pod \"ceilometer-0\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.722131 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/385f189a-9a2f-42db-a4c1-dd7b332a4f91-run-httpd\") pod \"ceilometer-0\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.722207 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-config-data\") pod \"ceilometer-0\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.722265 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.722325 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.722397 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/385f189a-9a2f-42db-a4c1-dd7b332a4f91-log-httpd\") pod \"ceilometer-0\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: E1013 09:03:27.727035 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825\": container with ID starting with 014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825 not found: ID does not exist" containerID="014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.727065 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825"} err="failed to get container status \"014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825\": rpc error: code = NotFound desc = could not find container \"014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825\": container with ID starting with 014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825 not found: ID does not exist" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.727083 4685 scope.go:117] "RemoveContainer" containerID="667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.727389 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2"} err="failed to get container status \"667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2\": rpc error: code = NotFound desc = could not find container \"667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2\": container with ID starting with 667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2 not found: ID does not exist" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.727434 4685 scope.go:117] "RemoveContainer" containerID="5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.727699 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4"} err="failed to get container status \"5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4\": rpc error: code = NotFound desc = could not find container \"5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4\": container with ID starting with 5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4 not found: ID does not exist" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.727762 4685 scope.go:117] "RemoveContainer" containerID="67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.728075 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8"} err="failed to get container status \"67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8\": rpc error: code = NotFound desc = could not find container \"67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8\": container with ID starting with 67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8 not found: ID does not exist" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.728095 4685 scope.go:117] "RemoveContainer" containerID="014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.729281 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825"} err="failed to get container status \"014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825\": rpc error: code = NotFound desc = could not find container \"014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825\": container with ID starting with 014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825 not found: ID does not exist" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.729343 4685 scope.go:117] "RemoveContainer" containerID="667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.734052 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2"} err="failed to get container status \"667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2\": rpc error: code = NotFound desc = could not find container \"667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2\": container with ID starting with 667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2 not found: ID does not exist" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.734089 4685 scope.go:117] "RemoveContainer" containerID="5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.734342 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4"} err="failed to get container status \"5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4\": rpc error: code = NotFound desc = could not find container \"5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4\": container with ID starting with 5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4 not found: ID does not exist" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.734365 4685 scope.go:117] "RemoveContainer" containerID="67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.734536 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8"} err="failed to get container status \"67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8\": rpc error: code = NotFound desc = could not find container \"67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8\": container with ID starting with 67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8 not found: ID does not exist" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.734560 4685 scope.go:117] "RemoveContainer" containerID="014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.734734 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825"} err="failed to get container status \"014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825\": rpc error: code = NotFound desc = could not find container \"014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825\": container with ID starting with 014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825 not found: ID does not exist" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.734757 4685 scope.go:117] "RemoveContainer" containerID="667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.734935 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2"} err="failed to get container status \"667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2\": rpc error: code = NotFound desc = could not find container \"667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2\": container with ID starting with 667aede49306936ff61b46568fa91fbbbdfbf9e4b5191d35561aa102f6dca1a2 not found: ID does not exist" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.734955 4685 scope.go:117] "RemoveContainer" containerID="5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.735377 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4"} err="failed to get container status \"5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4\": rpc error: code = NotFound desc = could not find container \"5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4\": container with ID starting with 5b966e2203568f686dd0ebd3d5123da9264e853c55fec501d17f5c32dfdb26c4 not found: ID does not exist" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.735395 4685 scope.go:117] "RemoveContainer" containerID="67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.736892 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8"} err="failed to get container status \"67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8\": rpc error: code = NotFound desc = could not find container \"67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8\": container with ID starting with 67d862909f0a18cb41adaddce9c6ff365c4f21426bfa2ccab826a4f47d02c2e8 not found: ID does not exist" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.736928 4685 scope.go:117] "RemoveContainer" containerID="014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.737107 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825"} err="failed to get container status \"014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825\": rpc error: code = NotFound desc = could not find container \"014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825\": container with ID starting with 014922f1e807e37c69924c8bae7d729395288fc7d2d443083d6685d0d7850825 not found: ID does not exist" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.824478 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pvpb\" (UniqueName: \"kubernetes.io/projected/385f189a-9a2f-42db-a4c1-dd7b332a4f91-kube-api-access-5pvpb\") pod \"ceilometer-0\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.824549 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/385f189a-9a2f-42db-a4c1-dd7b332a4f91-run-httpd\") pod \"ceilometer-0\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.824604 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-config-data\") pod \"ceilometer-0\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.824638 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.824682 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.824720 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/385f189a-9a2f-42db-a4c1-dd7b332a4f91-log-httpd\") pod \"ceilometer-0\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.824858 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-scripts\") pod \"ceilometer-0\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.825735 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/385f189a-9a2f-42db-a4c1-dd7b332a4f91-log-httpd\") pod \"ceilometer-0\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.828198 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/385f189a-9a2f-42db-a4c1-dd7b332a4f91-run-httpd\") pod \"ceilometer-0\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.832119 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.832179 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-config-data\") pod \"ceilometer-0\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.832605 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-scripts\") pod \"ceilometer-0\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.837651 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.851424 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pvpb\" (UniqueName: \"kubernetes.io/projected/385f189a-9a2f-42db-a4c1-dd7b332a4f91-kube-api-access-5pvpb\") pod \"ceilometer-0\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " pod="openstack/ceilometer-0" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.932671 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-cf2b-account-create-d9lnd"] Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.933837 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-cf2b-account-create-d9lnd" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.937709 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.946993 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-cf2b-account-create-d9lnd"] Oct 13 09:03:27 crc kubenswrapper[4685]: I1013 09:03:27.977702 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.027978 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7dwg\" (UniqueName: \"kubernetes.io/projected/a2581060-597a-46b8-861f-ff53d55748fc-kube-api-access-n7dwg\") pod \"nova-api-cf2b-account-create-d9lnd\" (UID: \"a2581060-597a-46b8-861f-ff53d55748fc\") " pod="openstack/nova-api-cf2b-account-create-d9lnd" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.130751 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7dwg\" (UniqueName: \"kubernetes.io/projected/a2581060-597a-46b8-861f-ff53d55748fc-kube-api-access-n7dwg\") pod \"nova-api-cf2b-account-create-d9lnd\" (UID: \"a2581060-597a-46b8-861f-ff53d55748fc\") " pod="openstack/nova-api-cf2b-account-create-d9lnd" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.146833 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-379c-account-create-z5kgm"] Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.161356 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7dwg\" (UniqueName: \"kubernetes.io/projected/a2581060-597a-46b8-861f-ff53d55748fc-kube-api-access-n7dwg\") pod \"nova-api-cf2b-account-create-d9lnd\" (UID: \"a2581060-597a-46b8-861f-ff53d55748fc\") " pod="openstack/nova-api-cf2b-account-create-d9lnd" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.175888 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-379c-account-create-z5kgm"] Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.176027 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-379c-account-create-z5kgm" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.178547 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.220687 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-db976b79d-q446d" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.232805 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf8s7\" (UniqueName: \"kubernetes.io/projected/fe673980-2c45-45d0-a724-0b66c6f3e5ee-kube-api-access-hf8s7\") pod \"nova-cell0-379c-account-create-z5kgm\" (UID: \"fe673980-2c45-45d0-a724-0b66c6f3e5ee\") " pod="openstack/nova-cell0-379c-account-create-z5kgm" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.275411 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-cf2b-account-create-d9lnd" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.318856 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-b46c-account-create-2rssp"] Oct 13 09:03:28 crc kubenswrapper[4685]: E1013 09:03:28.319689 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" containerName="horizon" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.319701 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" containerName="horizon" Oct 13 09:03:28 crc kubenswrapper[4685]: E1013 09:03:28.319716 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" containerName="horizon-log" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.319723 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" containerName="horizon-log" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.320190 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" containerName="horizon" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.320207 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" containerName="horizon-log" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.320877 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b46c-account-create-2rssp" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.325147 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.335171 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-config-data\") pod \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.335234 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-scripts\") pod \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.335268 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-horizon-tls-certs\") pod \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.335337 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-logs\") pod \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.335394 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-combined-ca-bundle\") pod \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.335439 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-horizon-secret-key\") pod \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.335523 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54czh\" (UniqueName: \"kubernetes.io/projected/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-kube-api-access-54czh\") pod \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\" (UID: \"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc\") " Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.335786 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf8s7\" (UniqueName: \"kubernetes.io/projected/fe673980-2c45-45d0-a724-0b66c6f3e5ee-kube-api-access-hf8s7\") pod \"nova-cell0-379c-account-create-z5kgm\" (UID: \"fe673980-2c45-45d0-a724-0b66c6f3e5ee\") " pod="openstack/nova-cell0-379c-account-create-z5kgm" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.336423 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-logs" (OuterVolumeSpecName: "logs") pod "2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" (UID: "2046bf00-14ee-4d8d-a7ac-ab6a211b93cc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.342213 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-b46c-account-create-2rssp"] Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.356890 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf8s7\" (UniqueName: \"kubernetes.io/projected/fe673980-2c45-45d0-a724-0b66c6f3e5ee-kube-api-access-hf8s7\") pod \"nova-cell0-379c-account-create-z5kgm\" (UID: \"fe673980-2c45-45d0-a724-0b66c6f3e5ee\") " pod="openstack/nova-cell0-379c-account-create-z5kgm" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.357161 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-kube-api-access-54czh" (OuterVolumeSpecName: "kube-api-access-54czh") pod "2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" (UID: "2046bf00-14ee-4d8d-a7ac-ab6a211b93cc"). InnerVolumeSpecName "kube-api-access-54czh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.358021 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" (UID: "2046bf00-14ee-4d8d-a7ac-ab6a211b93cc"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.379100 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-config-data" (OuterVolumeSpecName: "config-data") pod "2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" (UID: "2046bf00-14ee-4d8d-a7ac-ab6a211b93cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.379219 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" (UID: "2046bf00-14ee-4d8d-a7ac-ab6a211b93cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.408463 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-scripts" (OuterVolumeSpecName: "scripts") pod "2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" (UID: "2046bf00-14ee-4d8d-a7ac-ab6a211b93cc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.423655 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" (UID: "2046bf00-14ee-4d8d-a7ac-ab6a211b93cc"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.438647 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5m2x\" (UniqueName: \"kubernetes.io/projected/998d6585-7a36-44b8-a339-a7c8409df426-kube-api-access-p5m2x\") pod \"nova-cell1-b46c-account-create-2rssp\" (UID: \"998d6585-7a36-44b8-a339-a7c8409df426\") " pod="openstack/nova-cell1-b46c-account-create-2rssp" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.438805 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54czh\" (UniqueName: \"kubernetes.io/projected/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-kube-api-access-54czh\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.438820 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.438831 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.438840 4685 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.438848 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-logs\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.438855 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.438864 4685 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.508271 4685 generic.go:334] "Generic (PLEG): container finished" podID="2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" containerID="53e463eff1a6fb595b65a78ef21731ec4ee12f240e505e924644e92f36c5e9c4" exitCode=137 Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.508370 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-db976b79d-q446d" event={"ID":"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc","Type":"ContainerDied","Data":"53e463eff1a6fb595b65a78ef21731ec4ee12f240e505e924644e92f36c5e9c4"} Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.508397 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-db976b79d-q446d" event={"ID":"2046bf00-14ee-4d8d-a7ac-ab6a211b93cc","Type":"ContainerDied","Data":"2a683f973e75bb1e5041098aa03edac5179ea0a65e33a3ad0f90f6afc26d2f2d"} Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.508413 4685 scope.go:117] "RemoveContainer" containerID="87642f946b3e4dd470d2686255f2a7452ddf07f367a7a524a315e59e389b70f5" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.508528 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-db976b79d-q446d" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.512832 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-379c-account-create-z5kgm" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.520012 4685 generic.go:334] "Generic (PLEG): container finished" podID="babeefc6-9cbe-4f6c-9487-731a9270c1c4" containerID="3346630c966f3d7aff2a38529b3649c2ce0e925ee871b7630194486f5d391a51" exitCode=0 Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.520076 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"babeefc6-9cbe-4f6c-9487-731a9270c1c4","Type":"ContainerDied","Data":"3346630c966f3d7aff2a38529b3649c2ce0e925ee871b7630194486f5d391a51"} Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.540293 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5m2x\" (UniqueName: \"kubernetes.io/projected/998d6585-7a36-44b8-a339-a7c8409df426-kube-api-access-p5m2x\") pod \"nova-cell1-b46c-account-create-2rssp\" (UID: \"998d6585-7a36-44b8-a339-a7c8409df426\") " pod="openstack/nova-cell1-b46c-account-create-2rssp" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.597294 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5m2x\" (UniqueName: \"kubernetes.io/projected/998d6585-7a36-44b8-a339-a7c8409df426-kube-api-access-p5m2x\") pod \"nova-cell1-b46c-account-create-2rssp\" (UID: \"998d6585-7a36-44b8-a339-a7c8409df426\") " pod="openstack/nova-cell1-b46c-account-create-2rssp" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.600624 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-db976b79d-q446d"] Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.611034 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-db976b79d-q446d"] Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.646279 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.679603 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b46c-account-create-2rssp" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.766073 4685 scope.go:117] "RemoveContainer" containerID="53e463eff1a6fb595b65a78ef21731ec4ee12f240e505e924644e92f36c5e9c4" Oct 13 09:03:28 crc kubenswrapper[4685]: I1013 09:03:28.841650 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-cf2b-account-create-d9lnd"] Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.043314 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.057044 4685 scope.go:117] "RemoveContainer" containerID="87642f946b3e4dd470d2686255f2a7452ddf07f367a7a524a315e59e389b70f5" Oct 13 09:03:29 crc kubenswrapper[4685]: E1013 09:03:29.072815 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87642f946b3e4dd470d2686255f2a7452ddf07f367a7a524a315e59e389b70f5\": container with ID starting with 87642f946b3e4dd470d2686255f2a7452ddf07f367a7a524a315e59e389b70f5 not found: ID does not exist" containerID="87642f946b3e4dd470d2686255f2a7452ddf07f367a7a524a315e59e389b70f5" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.072854 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87642f946b3e4dd470d2686255f2a7452ddf07f367a7a524a315e59e389b70f5"} err="failed to get container status \"87642f946b3e4dd470d2686255f2a7452ddf07f367a7a524a315e59e389b70f5\": rpc error: code = NotFound desc = could not find container \"87642f946b3e4dd470d2686255f2a7452ddf07f367a7a524a315e59e389b70f5\": container with ID starting with 87642f946b3e4dd470d2686255f2a7452ddf07f367a7a524a315e59e389b70f5 not found: ID does not exist" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.072880 4685 scope.go:117] "RemoveContainer" containerID="53e463eff1a6fb595b65a78ef21731ec4ee12f240e505e924644e92f36c5e9c4" Oct 13 09:03:29 crc kubenswrapper[4685]: E1013 09:03:29.083984 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53e463eff1a6fb595b65a78ef21731ec4ee12f240e505e924644e92f36c5e9c4\": container with ID starting with 53e463eff1a6fb595b65a78ef21731ec4ee12f240e505e924644e92f36c5e9c4 not found: ID does not exist" containerID="53e463eff1a6fb595b65a78ef21731ec4ee12f240e505e924644e92f36c5e9c4" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.084686 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53e463eff1a6fb595b65a78ef21731ec4ee12f240e505e924644e92f36c5e9c4"} err="failed to get container status \"53e463eff1a6fb595b65a78ef21731ec4ee12f240e505e924644e92f36c5e9c4\": rpc error: code = NotFound desc = could not find container \"53e463eff1a6fb595b65a78ef21731ec4ee12f240e505e924644e92f36c5e9c4\": container with ID starting with 53e463eff1a6fb595b65a78ef21731ec4ee12f240e505e924644e92f36c5e9c4 not found: ID does not exist" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.095134 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-379c-account-create-z5kgm"] Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.152238 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-config-data\") pod \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.152294 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.152365 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/babeefc6-9cbe-4f6c-9487-731a9270c1c4-logs\") pod \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.152399 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-internal-tls-certs\") pod \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.152447 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpnbw\" (UniqueName: \"kubernetes.io/projected/babeefc6-9cbe-4f6c-9487-731a9270c1c4-kube-api-access-tpnbw\") pod \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.152465 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-scripts\") pod \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.152501 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-combined-ca-bundle\") pod \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.152601 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/babeefc6-9cbe-4f6c-9487-731a9270c1c4-httpd-run\") pod \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\" (UID: \"babeefc6-9cbe-4f6c-9487-731a9270c1c4\") " Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.153810 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/babeefc6-9cbe-4f6c-9487-731a9270c1c4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "babeefc6-9cbe-4f6c-9487-731a9270c1c4" (UID: "babeefc6-9cbe-4f6c-9487-731a9270c1c4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.154069 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/babeefc6-9cbe-4f6c-9487-731a9270c1c4-logs" (OuterVolumeSpecName: "logs") pod "babeefc6-9cbe-4f6c-9487-731a9270c1c4" (UID: "babeefc6-9cbe-4f6c-9487-731a9270c1c4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.157414 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "babeefc6-9cbe-4f6c-9487-731a9270c1c4" (UID: "babeefc6-9cbe-4f6c-9487-731a9270c1c4"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.160052 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/babeefc6-9cbe-4f6c-9487-731a9270c1c4-kube-api-access-tpnbw" (OuterVolumeSpecName: "kube-api-access-tpnbw") pod "babeefc6-9cbe-4f6c-9487-731a9270c1c4" (UID: "babeefc6-9cbe-4f6c-9487-731a9270c1c4"). InnerVolumeSpecName "kube-api-access-tpnbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.181175 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-scripts" (OuterVolumeSpecName: "scripts") pod "babeefc6-9cbe-4f6c-9487-731a9270c1c4" (UID: "babeefc6-9cbe-4f6c-9487-731a9270c1c4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.192818 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "babeefc6-9cbe-4f6c-9487-731a9270c1c4" (UID: "babeefc6-9cbe-4f6c-9487-731a9270c1c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.255493 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/babeefc6-9cbe-4f6c-9487-731a9270c1c4-logs\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.255534 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpnbw\" (UniqueName: \"kubernetes.io/projected/babeefc6-9cbe-4f6c-9487-731a9270c1c4-kube-api-access-tpnbw\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.255547 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.255559 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.255598 4685 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/babeefc6-9cbe-4f6c-9487-731a9270c1c4-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.255625 4685 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.288839 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "babeefc6-9cbe-4f6c-9487-731a9270c1c4" (UID: "babeefc6-9cbe-4f6c-9487-731a9270c1c4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.288998 4685 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.291600 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-config-data" (OuterVolumeSpecName: "config-data") pod "babeefc6-9cbe-4f6c-9487-731a9270c1c4" (UID: "babeefc6-9cbe-4f6c-9487-731a9270c1c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.357776 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.357821 4685 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.357830 4685 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/babeefc6-9cbe-4f6c-9487-731a9270c1c4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.465675 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-b46c-account-create-2rssp"] Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.514825 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ea0f496-a1c1-41e9-9b72-d72a9acecb02" path="/var/lib/kubelet/pods/1ea0f496-a1c1-41e9-9b72-d72a9acecb02/volumes" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.515743 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2046bf00-14ee-4d8d-a7ac-ab6a211b93cc" path="/var/lib/kubelet/pods/2046bf00-14ee-4d8d-a7ac-ab6a211b93cc/volumes" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.532525 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-379c-account-create-z5kgm" event={"ID":"fe673980-2c45-45d0-a724-0b66c6f3e5ee","Type":"ContainerStarted","Data":"9a43e73ccd84230ab7092c33005509c9cafefcf925de0862ed3f543176080c7b"} Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.533691 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b46c-account-create-2rssp" event={"ID":"998d6585-7a36-44b8-a339-a7c8409df426","Type":"ContainerStarted","Data":"5ec69e15379779577850836ddafeaab2326e02c0784b06b6c6783489445b02e5"} Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.540833 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"babeefc6-9cbe-4f6c-9487-731a9270c1c4","Type":"ContainerDied","Data":"dcc4502a4608563d57e3b8098244b876ac27b610f2a0f6a09abd8229da1f5935"} Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.540881 4685 scope.go:117] "RemoveContainer" containerID="3346630c966f3d7aff2a38529b3649c2ce0e925ee871b7630194486f5d391a51" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.541004 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.543311 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"385f189a-9a2f-42db-a4c1-dd7b332a4f91","Type":"ContainerStarted","Data":"a4e5db4e63dd41bb0ae291dc2f02d90b9c8ee5238553b86dd2b87c595cc51ff2"} Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.548840 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-cf2b-account-create-d9lnd" event={"ID":"a2581060-597a-46b8-861f-ff53d55748fc","Type":"ContainerStarted","Data":"42849fcc4860677169b406ff8ebb2bc1b98512daa12f4979ad2ec2e0f904b99d"} Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.591578 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.601064 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.610785 4685 scope.go:117] "RemoveContainer" containerID="1151b5de139a4c03b47ac9c2ee15276efe6b1481c2f0cba8e9123177807e5804" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.612253 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 09:03:29 crc kubenswrapper[4685]: E1013 09:03:29.612696 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="babeefc6-9cbe-4f6c-9487-731a9270c1c4" containerName="glance-httpd" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.612712 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="babeefc6-9cbe-4f6c-9487-731a9270c1c4" containerName="glance-httpd" Oct 13 09:03:29 crc kubenswrapper[4685]: E1013 09:03:29.612730 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="babeefc6-9cbe-4f6c-9487-731a9270c1c4" containerName="glance-log" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.612737 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="babeefc6-9cbe-4f6c-9487-731a9270c1c4" containerName="glance-log" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.612959 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="babeefc6-9cbe-4f6c-9487-731a9270c1c4" containerName="glance-log" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.612986 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="babeefc6-9cbe-4f6c-9487-731a9270c1c4" containerName="glance-httpd" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.613859 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.625132 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.625346 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.646046 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.673579 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96ceb8f4-f3db-4f87-b5bf-27fd14076c1b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.673633 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/96ceb8f4-f3db-4f87-b5bf-27fd14076c1b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.673658 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7q84\" (UniqueName: \"kubernetes.io/projected/96ceb8f4-f3db-4f87-b5bf-27fd14076c1b-kube-api-access-b7q84\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.673716 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96ceb8f4-f3db-4f87-b5bf-27fd14076c1b-logs\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.673744 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.673854 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/96ceb8f4-f3db-4f87-b5bf-27fd14076c1b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.674020 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96ceb8f4-f3db-4f87-b5bf-27fd14076c1b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.674110 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96ceb8f4-f3db-4f87-b5bf-27fd14076c1b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.776847 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96ceb8f4-f3db-4f87-b5bf-27fd14076c1b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.776888 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/96ceb8f4-f3db-4f87-b5bf-27fd14076c1b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.776925 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7q84\" (UniqueName: \"kubernetes.io/projected/96ceb8f4-f3db-4f87-b5bf-27fd14076c1b-kube-api-access-b7q84\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.776957 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96ceb8f4-f3db-4f87-b5bf-27fd14076c1b-logs\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.776978 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.777034 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/96ceb8f4-f3db-4f87-b5bf-27fd14076c1b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.777074 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96ceb8f4-f3db-4f87-b5bf-27fd14076c1b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.777102 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96ceb8f4-f3db-4f87-b5bf-27fd14076c1b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.778002 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.778159 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/96ceb8f4-f3db-4f87-b5bf-27fd14076c1b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.778376 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96ceb8f4-f3db-4f87-b5bf-27fd14076c1b-logs\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.783789 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96ceb8f4-f3db-4f87-b5bf-27fd14076c1b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.784291 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/96ceb8f4-f3db-4f87-b5bf-27fd14076c1b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.785331 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96ceb8f4-f3db-4f87-b5bf-27fd14076c1b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.785949 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96ceb8f4-f3db-4f87-b5bf-27fd14076c1b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.796748 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7q84\" (UniqueName: \"kubernetes.io/projected/96ceb8f4-f3db-4f87-b5bf-27fd14076c1b-kube-api-access-b7q84\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.816138 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b\") " pod="openstack/glance-default-internal-api-0" Oct 13 09:03:29 crc kubenswrapper[4685]: I1013 09:03:29.979760 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 09:03:30 crc kubenswrapper[4685]: I1013 09:03:30.567558 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-cf2b-account-create-d9lnd" event={"ID":"a2581060-597a-46b8-861f-ff53d55748fc","Type":"ContainerStarted","Data":"94fa93c93a9ce7c794e14c1fbcc577fb3bd0ffa09823ea05fdf9f6ca44d188c4"} Oct 13 09:03:30 crc kubenswrapper[4685]: I1013 09:03:30.569566 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-379c-account-create-z5kgm" event={"ID":"fe673980-2c45-45d0-a724-0b66c6f3e5ee","Type":"ContainerStarted","Data":"e6a919343f2e0025221223526df6f1e6f23a7bf2d7bb5ecd786fe39aed46ab3c"} Oct 13 09:03:30 crc kubenswrapper[4685]: I1013 09:03:30.571367 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b46c-account-create-2rssp" event={"ID":"998d6585-7a36-44b8-a339-a7c8409df426","Type":"ContainerStarted","Data":"c1b8da12450a58ae7a4ce71b7afc3b047747dc08dd9165b1c8a13d4c1f03ee0d"} Oct 13 09:03:30 crc kubenswrapper[4685]: I1013 09:03:30.606186 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-cf2b-account-create-d9lnd" podStartSLOduration=3.606167716 podStartE2EDuration="3.606167716s" podCreationTimestamp="2025-10-13 09:03:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:03:30.588869745 +0000 UTC m=+1135.736745506" watchObservedRunningTime="2025-10-13 09:03:30.606167716 +0000 UTC m=+1135.754043477" Oct 13 09:03:30 crc kubenswrapper[4685]: I1013 09:03:30.609838 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-379c-account-create-z5kgm" podStartSLOduration=2.609825496 podStartE2EDuration="2.609825496s" podCreationTimestamp="2025-10-13 09:03:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:03:30.602635809 +0000 UTC m=+1135.750511570" watchObservedRunningTime="2025-10-13 09:03:30.609825496 +0000 UTC m=+1135.757701257" Oct 13 09:03:30 crc kubenswrapper[4685]: I1013 09:03:30.652645 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-b46c-account-create-2rssp" podStartSLOduration=2.652624151 podStartE2EDuration="2.652624151s" podCreationTimestamp="2025-10-13 09:03:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:03:30.640578743 +0000 UTC m=+1135.788454504" watchObservedRunningTime="2025-10-13 09:03:30.652624151 +0000 UTC m=+1135.800499912" Oct 13 09:03:30 crc kubenswrapper[4685]: I1013 09:03:30.665421 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 09:03:31 crc kubenswrapper[4685]: I1013 09:03:31.525955 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="babeefc6-9cbe-4f6c-9487-731a9270c1c4" path="/var/lib/kubelet/pods/babeefc6-9cbe-4f6c-9487-731a9270c1c4/volumes" Oct 13 09:03:31 crc kubenswrapper[4685]: I1013 09:03:31.583935 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"385f189a-9a2f-42db-a4c1-dd7b332a4f91","Type":"ContainerStarted","Data":"72ca217c6a84e6ba2a6a56c65d6b450815a3b82fa6b50e64cd15f08f1f71f625"} Oct 13 09:03:31 crc kubenswrapper[4685]: I1013 09:03:31.586851 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b","Type":"ContainerStarted","Data":"05d20b865c9171dba6e917bf215c2718fa92634009bbed78196120614ff7bb37"} Oct 13 09:03:31 crc kubenswrapper[4685]: I1013 09:03:31.586904 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b","Type":"ContainerStarted","Data":"14e628a299cc4ab1a4d46e5dcce69e1485163226039a321cdad5f725cd2ecea0"} Oct 13 09:03:31 crc kubenswrapper[4685]: I1013 09:03:31.589040 4685 generic.go:334] "Generic (PLEG): container finished" podID="a2581060-597a-46b8-861f-ff53d55748fc" containerID="94fa93c93a9ce7c794e14c1fbcc577fb3bd0ffa09823ea05fdf9f6ca44d188c4" exitCode=0 Oct 13 09:03:31 crc kubenswrapper[4685]: I1013 09:03:31.589098 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-cf2b-account-create-d9lnd" event={"ID":"a2581060-597a-46b8-861f-ff53d55748fc","Type":"ContainerDied","Data":"94fa93c93a9ce7c794e14c1fbcc577fb3bd0ffa09823ea05fdf9f6ca44d188c4"} Oct 13 09:03:31 crc kubenswrapper[4685]: I1013 09:03:31.592168 4685 generic.go:334] "Generic (PLEG): container finished" podID="fe673980-2c45-45d0-a724-0b66c6f3e5ee" containerID="e6a919343f2e0025221223526df6f1e6f23a7bf2d7bb5ecd786fe39aed46ab3c" exitCode=0 Oct 13 09:03:31 crc kubenswrapper[4685]: I1013 09:03:31.592223 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-379c-account-create-z5kgm" event={"ID":"fe673980-2c45-45d0-a724-0b66c6f3e5ee","Type":"ContainerDied","Data":"e6a919343f2e0025221223526df6f1e6f23a7bf2d7bb5ecd786fe39aed46ab3c"} Oct 13 09:03:31 crc kubenswrapper[4685]: I1013 09:03:31.601100 4685 generic.go:334] "Generic (PLEG): container finished" podID="998d6585-7a36-44b8-a339-a7c8409df426" containerID="c1b8da12450a58ae7a4ce71b7afc3b047747dc08dd9165b1c8a13d4c1f03ee0d" exitCode=0 Oct 13 09:03:31 crc kubenswrapper[4685]: I1013 09:03:31.601148 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b46c-account-create-2rssp" event={"ID":"998d6585-7a36-44b8-a339-a7c8409df426","Type":"ContainerDied","Data":"c1b8da12450a58ae7a4ce71b7afc3b047747dc08dd9165b1c8a13d4c1f03ee0d"} Oct 13 09:03:32 crc kubenswrapper[4685]: I1013 09:03:32.615465 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"385f189a-9a2f-42db-a4c1-dd7b332a4f91","Type":"ContainerStarted","Data":"e2c105cfd278ab4993562c530a09a53746a65cb680c3923dc6a1c7b55b03dc70"} Oct 13 09:03:32 crc kubenswrapper[4685]: I1013 09:03:32.623943 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"96ceb8f4-f3db-4f87-b5bf-27fd14076c1b","Type":"ContainerStarted","Data":"8e7e9fe35fae7a9e20fd0fbdfd5de69e6bd984cf5ba1c2fc9a4953859efc8a65"} Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.160441 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b46c-account-create-2rssp" Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.181140 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5m2x\" (UniqueName: \"kubernetes.io/projected/998d6585-7a36-44b8-a339-a7c8409df426-kube-api-access-p5m2x\") pod \"998d6585-7a36-44b8-a339-a7c8409df426\" (UID: \"998d6585-7a36-44b8-a339-a7c8409df426\") " Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.195079 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.195039098 podStartE2EDuration="4.195039098s" podCreationTimestamp="2025-10-13 09:03:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:03:32.664406119 +0000 UTC m=+1137.812281880" watchObservedRunningTime="2025-10-13 09:03:33.195039098 +0000 UTC m=+1138.342914859" Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.195210 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/998d6585-7a36-44b8-a339-a7c8409df426-kube-api-access-p5m2x" (OuterVolumeSpecName: "kube-api-access-p5m2x") pod "998d6585-7a36-44b8-a339-a7c8409df426" (UID: "998d6585-7a36-44b8-a339-a7c8409df426"). InnerVolumeSpecName "kube-api-access-p5m2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.199747 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-cf2b-account-create-d9lnd" Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.231493 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-379c-account-create-z5kgm" Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.282289 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7dwg\" (UniqueName: \"kubernetes.io/projected/a2581060-597a-46b8-861f-ff53d55748fc-kube-api-access-n7dwg\") pod \"a2581060-597a-46b8-861f-ff53d55748fc\" (UID: \"a2581060-597a-46b8-861f-ff53d55748fc\") " Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.282340 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hf8s7\" (UniqueName: \"kubernetes.io/projected/fe673980-2c45-45d0-a724-0b66c6f3e5ee-kube-api-access-hf8s7\") pod \"fe673980-2c45-45d0-a724-0b66c6f3e5ee\" (UID: \"fe673980-2c45-45d0-a724-0b66c6f3e5ee\") " Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.282664 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5m2x\" (UniqueName: \"kubernetes.io/projected/998d6585-7a36-44b8-a339-a7c8409df426-kube-api-access-p5m2x\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.287237 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2581060-597a-46b8-861f-ff53d55748fc-kube-api-access-n7dwg" (OuterVolumeSpecName: "kube-api-access-n7dwg") pod "a2581060-597a-46b8-861f-ff53d55748fc" (UID: "a2581060-597a-46b8-861f-ff53d55748fc"). InnerVolumeSpecName "kube-api-access-n7dwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.289050 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe673980-2c45-45d0-a724-0b66c6f3e5ee-kube-api-access-hf8s7" (OuterVolumeSpecName: "kube-api-access-hf8s7") pod "fe673980-2c45-45d0-a724-0b66c6f3e5ee" (UID: "fe673980-2c45-45d0-a724-0b66c6f3e5ee"). InnerVolumeSpecName "kube-api-access-hf8s7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.383805 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7dwg\" (UniqueName: \"kubernetes.io/projected/a2581060-597a-46b8-861f-ff53d55748fc-kube-api-access-n7dwg\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.383853 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hf8s7\" (UniqueName: \"kubernetes.io/projected/fe673980-2c45-45d0-a724-0b66c6f3e5ee-kube-api-access-hf8s7\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.639644 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b46c-account-create-2rssp" event={"ID":"998d6585-7a36-44b8-a339-a7c8409df426","Type":"ContainerDied","Data":"5ec69e15379779577850836ddafeaab2326e02c0784b06b6c6783489445b02e5"} Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.639683 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ec69e15379779577850836ddafeaab2326e02c0784b06b6c6783489445b02e5" Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.639693 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b46c-account-create-2rssp" Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.644550 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"385f189a-9a2f-42db-a4c1-dd7b332a4f91","Type":"ContainerStarted","Data":"7f2459f6bdb04ed6437aa0a38c51919274e41eb2c9c2bed408f9f44221a1cf0a"} Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.646541 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-cf2b-account-create-d9lnd" event={"ID":"a2581060-597a-46b8-861f-ff53d55748fc","Type":"ContainerDied","Data":"42849fcc4860677169b406ff8ebb2bc1b98512daa12f4979ad2ec2e0f904b99d"} Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.646582 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42849fcc4860677169b406ff8ebb2bc1b98512daa12f4979ad2ec2e0f904b99d" Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.646585 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-cf2b-account-create-d9lnd" Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.648888 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-379c-account-create-z5kgm" Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.649231 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-379c-account-create-z5kgm" event={"ID":"fe673980-2c45-45d0-a724-0b66c6f3e5ee","Type":"ContainerDied","Data":"9a43e73ccd84230ab7092c33005509c9cafefcf925de0862ed3f543176080c7b"} Oct 13 09:03:33 crc kubenswrapper[4685]: I1013 09:03:33.649285 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a43e73ccd84230ab7092c33005509c9cafefcf925de0862ed3f543176080c7b" Oct 13 09:03:34 crc kubenswrapper[4685]: I1013 09:03:34.326291 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 09:03:34 crc kubenswrapper[4685]: I1013 09:03:34.327149 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e81aae5c-c9f2-455c-8154-fa64d948bcab" containerName="glance-log" containerID="cri-o://5410b128f2e6ae710b0ee1323e2c57e92c1d6fee98ce55182f7f60d4a792dee0" gracePeriod=30 Oct 13 09:03:34 crc kubenswrapper[4685]: I1013 09:03:34.327324 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e81aae5c-c9f2-455c-8154-fa64d948bcab" containerName="glance-httpd" containerID="cri-o://aba2a67b03bcffa3d89c46a94c08cfd4f235737c2fabffcb870c7a5181e0fbde" gracePeriod=30 Oct 13 09:03:34 crc kubenswrapper[4685]: I1013 09:03:34.658797 4685 generic.go:334] "Generic (PLEG): container finished" podID="e81aae5c-c9f2-455c-8154-fa64d948bcab" containerID="5410b128f2e6ae710b0ee1323e2c57e92c1d6fee98ce55182f7f60d4a792dee0" exitCode=143 Oct 13 09:03:34 crc kubenswrapper[4685]: I1013 09:03:34.658835 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e81aae5c-c9f2-455c-8154-fa64d948bcab","Type":"ContainerDied","Data":"5410b128f2e6ae710b0ee1323e2c57e92c1d6fee98ce55182f7f60d4a792dee0"} Oct 13 09:03:35 crc kubenswrapper[4685]: I1013 09:03:35.309064 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 13 09:03:36 crc kubenswrapper[4685]: I1013 09:03:36.676055 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"385f189a-9a2f-42db-a4c1-dd7b332a4f91","Type":"ContainerStarted","Data":"0c1c949c42fc6a5194f10fa430e54a5f7fe1b292d7c4a9ea5b54dc5e2f8906e9"} Oct 13 09:03:36 crc kubenswrapper[4685]: I1013 09:03:36.677058 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 13 09:03:36 crc kubenswrapper[4685]: I1013 09:03:36.704266 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.692271381 podStartE2EDuration="9.704248309s" podCreationTimestamp="2025-10-13 09:03:27 +0000 UTC" firstStartedPulling="2025-10-13 09:03:28.803202223 +0000 UTC m=+1133.951077984" lastFinishedPulling="2025-10-13 09:03:35.815179151 +0000 UTC m=+1140.963054912" observedRunningTime="2025-10-13 09:03:36.695372968 +0000 UTC m=+1141.843248729" watchObservedRunningTime="2025-10-13 09:03:36.704248309 +0000 UTC m=+1141.852124070" Oct 13 09:03:37 crc kubenswrapper[4685]: I1013 09:03:37.685646 4685 generic.go:334] "Generic (PLEG): container finished" podID="e81aae5c-c9f2-455c-8154-fa64d948bcab" containerID="aba2a67b03bcffa3d89c46a94c08cfd4f235737c2fabffcb870c7a5181e0fbde" exitCode=0 Oct 13 09:03:37 crc kubenswrapper[4685]: I1013 09:03:37.685689 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e81aae5c-c9f2-455c-8154-fa64d948bcab","Type":"ContainerDied","Data":"aba2a67b03bcffa3d89c46a94c08cfd4f235737c2fabffcb870c7a5181e0fbde"} Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.121823 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.282014 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-public-tls-certs\") pod \"e81aae5c-c9f2-455c-8154-fa64d948bcab\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.282114 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n2jc\" (UniqueName: \"kubernetes.io/projected/e81aae5c-c9f2-455c-8154-fa64d948bcab-kube-api-access-9n2jc\") pod \"e81aae5c-c9f2-455c-8154-fa64d948bcab\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.282139 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-config-data\") pod \"e81aae5c-c9f2-455c-8154-fa64d948bcab\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.282174 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"e81aae5c-c9f2-455c-8154-fa64d948bcab\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.282196 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-combined-ca-bundle\") pod \"e81aae5c-c9f2-455c-8154-fa64d948bcab\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.282215 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e81aae5c-c9f2-455c-8154-fa64d948bcab-httpd-run\") pod \"e81aae5c-c9f2-455c-8154-fa64d948bcab\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.282268 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e81aae5c-c9f2-455c-8154-fa64d948bcab-logs\") pod \"e81aae5c-c9f2-455c-8154-fa64d948bcab\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.282352 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-scripts\") pod \"e81aae5c-c9f2-455c-8154-fa64d948bcab\" (UID: \"e81aae5c-c9f2-455c-8154-fa64d948bcab\") " Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.283244 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e81aae5c-c9f2-455c-8154-fa64d948bcab-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e81aae5c-c9f2-455c-8154-fa64d948bcab" (UID: "e81aae5c-c9f2-455c-8154-fa64d948bcab"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.283286 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e81aae5c-c9f2-455c-8154-fa64d948bcab-logs" (OuterVolumeSpecName: "logs") pod "e81aae5c-c9f2-455c-8154-fa64d948bcab" (UID: "e81aae5c-c9f2-455c-8154-fa64d948bcab"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.287544 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "e81aae5c-c9f2-455c-8154-fa64d948bcab" (UID: "e81aae5c-c9f2-455c-8154-fa64d948bcab"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.287870 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e81aae5c-c9f2-455c-8154-fa64d948bcab-kube-api-access-9n2jc" (OuterVolumeSpecName: "kube-api-access-9n2jc") pod "e81aae5c-c9f2-455c-8154-fa64d948bcab" (UID: "e81aae5c-c9f2-455c-8154-fa64d948bcab"). InnerVolumeSpecName "kube-api-access-9n2jc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.320040 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e81aae5c-c9f2-455c-8154-fa64d948bcab" (UID: "e81aae5c-c9f2-455c-8154-fa64d948bcab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.321731 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-scripts" (OuterVolumeSpecName: "scripts") pod "e81aae5c-c9f2-455c-8154-fa64d948bcab" (UID: "e81aae5c-c9f2-455c-8154-fa64d948bcab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.347413 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e81aae5c-c9f2-455c-8154-fa64d948bcab" (UID: "e81aae5c-c9f2-455c-8154-fa64d948bcab"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.387058 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.387096 4685 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.387109 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n2jc\" (UniqueName: \"kubernetes.io/projected/e81aae5c-c9f2-455c-8154-fa64d948bcab-kube-api-access-9n2jc\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.387153 4685 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.387169 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.387182 4685 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e81aae5c-c9f2-455c-8154-fa64d948bcab-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.387194 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e81aae5c-c9f2-455c-8154-fa64d948bcab-logs\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.422056 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-config-data" (OuterVolumeSpecName: "config-data") pod "e81aae5c-c9f2-455c-8154-fa64d948bcab" (UID: "e81aae5c-c9f2-455c-8154-fa64d948bcab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.437146 4685 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.457318 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-l4jzf"] Oct 13 09:03:38 crc kubenswrapper[4685]: E1013 09:03:38.457741 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e81aae5c-c9f2-455c-8154-fa64d948bcab" containerName="glance-httpd" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.457760 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e81aae5c-c9f2-455c-8154-fa64d948bcab" containerName="glance-httpd" Oct 13 09:03:38 crc kubenswrapper[4685]: E1013 09:03:38.457780 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e81aae5c-c9f2-455c-8154-fa64d948bcab" containerName="glance-log" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.457787 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e81aae5c-c9f2-455c-8154-fa64d948bcab" containerName="glance-log" Oct 13 09:03:38 crc kubenswrapper[4685]: E1013 09:03:38.457796 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2581060-597a-46b8-861f-ff53d55748fc" containerName="mariadb-account-create" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.457803 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2581060-597a-46b8-861f-ff53d55748fc" containerName="mariadb-account-create" Oct 13 09:03:38 crc kubenswrapper[4685]: E1013 09:03:38.457823 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="998d6585-7a36-44b8-a339-a7c8409df426" containerName="mariadb-account-create" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.457830 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="998d6585-7a36-44b8-a339-a7c8409df426" containerName="mariadb-account-create" Oct 13 09:03:38 crc kubenswrapper[4685]: E1013 09:03:38.457845 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe673980-2c45-45d0-a724-0b66c6f3e5ee" containerName="mariadb-account-create" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.457852 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe673980-2c45-45d0-a724-0b66c6f3e5ee" containerName="mariadb-account-create" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.458131 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe673980-2c45-45d0-a724-0b66c6f3e5ee" containerName="mariadb-account-create" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.458151 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="e81aae5c-c9f2-455c-8154-fa64d948bcab" containerName="glance-log" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.458178 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="998d6585-7a36-44b8-a339-a7c8409df426" containerName="mariadb-account-create" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.458193 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="e81aae5c-c9f2-455c-8154-fa64d948bcab" containerName="glance-httpd" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.458213 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2581060-597a-46b8-861f-ff53d55748fc" containerName="mariadb-account-create" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.458864 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-l4jzf" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.461479 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.461628 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-lqtxf" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.461807 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.483562 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-l4jzf"] Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.488793 4685 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.488821 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e81aae5c-c9f2-455c-8154-fa64d948bcab-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.590653 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11c738eb-e9be-4e64-a38c-da231145e8ab-config-data\") pod \"nova-cell0-conductor-db-sync-l4jzf\" (UID: \"11c738eb-e9be-4e64-a38c-da231145e8ab\") " pod="openstack/nova-cell0-conductor-db-sync-l4jzf" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.590701 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11c738eb-e9be-4e64-a38c-da231145e8ab-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-l4jzf\" (UID: \"11c738eb-e9be-4e64-a38c-da231145e8ab\") " pod="openstack/nova-cell0-conductor-db-sync-l4jzf" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.590818 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11c738eb-e9be-4e64-a38c-da231145e8ab-scripts\") pod \"nova-cell0-conductor-db-sync-l4jzf\" (UID: \"11c738eb-e9be-4e64-a38c-da231145e8ab\") " pod="openstack/nova-cell0-conductor-db-sync-l4jzf" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.590899 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjmml\" (UniqueName: \"kubernetes.io/projected/11c738eb-e9be-4e64-a38c-da231145e8ab-kube-api-access-gjmml\") pod \"nova-cell0-conductor-db-sync-l4jzf\" (UID: \"11c738eb-e9be-4e64-a38c-da231145e8ab\") " pod="openstack/nova-cell0-conductor-db-sync-l4jzf" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.692616 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjmml\" (UniqueName: \"kubernetes.io/projected/11c738eb-e9be-4e64-a38c-da231145e8ab-kube-api-access-gjmml\") pod \"nova-cell0-conductor-db-sync-l4jzf\" (UID: \"11c738eb-e9be-4e64-a38c-da231145e8ab\") " pod="openstack/nova-cell0-conductor-db-sync-l4jzf" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.692692 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11c738eb-e9be-4e64-a38c-da231145e8ab-config-data\") pod \"nova-cell0-conductor-db-sync-l4jzf\" (UID: \"11c738eb-e9be-4e64-a38c-da231145e8ab\") " pod="openstack/nova-cell0-conductor-db-sync-l4jzf" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.692738 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11c738eb-e9be-4e64-a38c-da231145e8ab-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-l4jzf\" (UID: \"11c738eb-e9be-4e64-a38c-da231145e8ab\") " pod="openstack/nova-cell0-conductor-db-sync-l4jzf" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.692860 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11c738eb-e9be-4e64-a38c-da231145e8ab-scripts\") pod \"nova-cell0-conductor-db-sync-l4jzf\" (UID: \"11c738eb-e9be-4e64-a38c-da231145e8ab\") " pod="openstack/nova-cell0-conductor-db-sync-l4jzf" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.697722 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11c738eb-e9be-4e64-a38c-da231145e8ab-config-data\") pod \"nova-cell0-conductor-db-sync-l4jzf\" (UID: \"11c738eb-e9be-4e64-a38c-da231145e8ab\") " pod="openstack/nova-cell0-conductor-db-sync-l4jzf" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.703838 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11c738eb-e9be-4e64-a38c-da231145e8ab-scripts\") pod \"nova-cell0-conductor-db-sync-l4jzf\" (UID: \"11c738eb-e9be-4e64-a38c-da231145e8ab\") " pod="openstack/nova-cell0-conductor-db-sync-l4jzf" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.704173 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e81aae5c-c9f2-455c-8154-fa64d948bcab","Type":"ContainerDied","Data":"bf89c9a2a3f92e59f2152f7c8cad3c8ee995e7f42b49c371fa94b397af442c4f"} Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.704273 4685 scope.go:117] "RemoveContainer" containerID="aba2a67b03bcffa3d89c46a94c08cfd4f235737c2fabffcb870c7a5181e0fbde" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.704503 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.709621 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjmml\" (UniqueName: \"kubernetes.io/projected/11c738eb-e9be-4e64-a38c-da231145e8ab-kube-api-access-gjmml\") pod \"nova-cell0-conductor-db-sync-l4jzf\" (UID: \"11c738eb-e9be-4e64-a38c-da231145e8ab\") " pod="openstack/nova-cell0-conductor-db-sync-l4jzf" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.726577 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11c738eb-e9be-4e64-a38c-da231145e8ab-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-l4jzf\" (UID: \"11c738eb-e9be-4e64-a38c-da231145e8ab\") " pod="openstack/nova-cell0-conductor-db-sync-l4jzf" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.773978 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-l4jzf" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.802371 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.820456 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.831130 4685 scope.go:117] "RemoveContainer" containerID="5410b128f2e6ae710b0ee1323e2c57e92c1d6fee98ce55182f7f60d4a792dee0" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.848829 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.850706 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.855117 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.861953 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.890093 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.998146 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf7cb050-76b4-4ee0-b71f-3ad58070fdc3-logs\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.998375 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf7cb050-76b4-4ee0-b71f-3ad58070fdc3-config-data\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.998398 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf7cb050-76b4-4ee0-b71f-3ad58070fdc3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.998429 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.998465 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cf7cb050-76b4-4ee0-b71f-3ad58070fdc3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.998504 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf7cb050-76b4-4ee0-b71f-3ad58070fdc3-scripts\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.998567 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbtzx\" (UniqueName: \"kubernetes.io/projected/cf7cb050-76b4-4ee0-b71f-3ad58070fdc3-kube-api-access-rbtzx\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:38 crc kubenswrapper[4685]: I1013 09:03:38.998582 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf7cb050-76b4-4ee0-b71f-3ad58070fdc3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.099842 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbtzx\" (UniqueName: \"kubernetes.io/projected/cf7cb050-76b4-4ee0-b71f-3ad58070fdc3-kube-api-access-rbtzx\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.099874 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf7cb050-76b4-4ee0-b71f-3ad58070fdc3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.099939 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf7cb050-76b4-4ee0-b71f-3ad58070fdc3-logs\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.099961 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf7cb050-76b4-4ee0-b71f-3ad58070fdc3-config-data\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.099980 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf7cb050-76b4-4ee0-b71f-3ad58070fdc3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.100004 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.100037 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cf7cb050-76b4-4ee0-b71f-3ad58070fdc3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.100073 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf7cb050-76b4-4ee0-b71f-3ad58070fdc3-scripts\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.100604 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.101114 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf7cb050-76b4-4ee0-b71f-3ad58070fdc3-logs\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.101353 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cf7cb050-76b4-4ee0-b71f-3ad58070fdc3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.109067 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf7cb050-76b4-4ee0-b71f-3ad58070fdc3-scripts\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.113222 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf7cb050-76b4-4ee0-b71f-3ad58070fdc3-config-data\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.120795 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbtzx\" (UniqueName: \"kubernetes.io/projected/cf7cb050-76b4-4ee0-b71f-3ad58070fdc3-kube-api-access-rbtzx\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.121214 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf7cb050-76b4-4ee0-b71f-3ad58070fdc3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.121685 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf7cb050-76b4-4ee0-b71f-3ad58070fdc3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.160751 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3\") " pod="openstack/glance-default-external-api-0" Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.234910 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.422264 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-l4jzf"] Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.517332 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e81aae5c-c9f2-455c-8154-fa64d948bcab" path="/var/lib/kubelet/pods/e81aae5c-c9f2-455c-8154-fa64d948bcab/volumes" Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.713221 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-l4jzf" event={"ID":"11c738eb-e9be-4e64-a38c-da231145e8ab","Type":"ContainerStarted","Data":"319477a66fdf8aa6e29e69a73736be566414b31696baface8ea779451838061a"} Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.850780 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.980127 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 13 09:03:39 crc kubenswrapper[4685]: I1013 09:03:39.980385 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 13 09:03:40 crc kubenswrapper[4685]: I1013 09:03:40.101524 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:40 crc kubenswrapper[4685]: I1013 09:03:40.101778 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="385f189a-9a2f-42db-a4c1-dd7b332a4f91" containerName="ceilometer-central-agent" containerID="cri-o://72ca217c6a84e6ba2a6a56c65d6b450815a3b82fa6b50e64cd15f08f1f71f625" gracePeriod=30 Oct 13 09:03:40 crc kubenswrapper[4685]: I1013 09:03:40.102255 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="385f189a-9a2f-42db-a4c1-dd7b332a4f91" containerName="proxy-httpd" containerID="cri-o://0c1c949c42fc6a5194f10fa430e54a5f7fe1b292d7c4a9ea5b54dc5e2f8906e9" gracePeriod=30 Oct 13 09:03:40 crc kubenswrapper[4685]: I1013 09:03:40.102305 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="385f189a-9a2f-42db-a4c1-dd7b332a4f91" containerName="sg-core" containerID="cri-o://7f2459f6bdb04ed6437aa0a38c51919274e41eb2c9c2bed408f9f44221a1cf0a" gracePeriod=30 Oct 13 09:03:40 crc kubenswrapper[4685]: I1013 09:03:40.102339 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="385f189a-9a2f-42db-a4c1-dd7b332a4f91" containerName="ceilometer-notification-agent" containerID="cri-o://e2c105cfd278ab4993562c530a09a53746a65cb680c3923dc6a1c7b55b03dc70" gracePeriod=30 Oct 13 09:03:40 crc kubenswrapper[4685]: I1013 09:03:40.104585 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 13 09:03:40 crc kubenswrapper[4685]: I1013 09:03:40.158064 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 13 09:03:40 crc kubenswrapper[4685]: I1013 09:03:40.727566 4685 generic.go:334] "Generic (PLEG): container finished" podID="385f189a-9a2f-42db-a4c1-dd7b332a4f91" containerID="0c1c949c42fc6a5194f10fa430e54a5f7fe1b292d7c4a9ea5b54dc5e2f8906e9" exitCode=0 Oct 13 09:03:40 crc kubenswrapper[4685]: I1013 09:03:40.727595 4685 generic.go:334] "Generic (PLEG): container finished" podID="385f189a-9a2f-42db-a4c1-dd7b332a4f91" containerID="7f2459f6bdb04ed6437aa0a38c51919274e41eb2c9c2bed408f9f44221a1cf0a" exitCode=2 Oct 13 09:03:40 crc kubenswrapper[4685]: I1013 09:03:40.727603 4685 generic.go:334] "Generic (PLEG): container finished" podID="385f189a-9a2f-42db-a4c1-dd7b332a4f91" containerID="e2c105cfd278ab4993562c530a09a53746a65cb680c3923dc6a1c7b55b03dc70" exitCode=0 Oct 13 09:03:40 crc kubenswrapper[4685]: I1013 09:03:40.727636 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"385f189a-9a2f-42db-a4c1-dd7b332a4f91","Type":"ContainerDied","Data":"0c1c949c42fc6a5194f10fa430e54a5f7fe1b292d7c4a9ea5b54dc5e2f8906e9"} Oct 13 09:03:40 crc kubenswrapper[4685]: I1013 09:03:40.727662 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"385f189a-9a2f-42db-a4c1-dd7b332a4f91","Type":"ContainerDied","Data":"7f2459f6bdb04ed6437aa0a38c51919274e41eb2c9c2bed408f9f44221a1cf0a"} Oct 13 09:03:40 crc kubenswrapper[4685]: I1013 09:03:40.727671 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"385f189a-9a2f-42db-a4c1-dd7b332a4f91","Type":"ContainerDied","Data":"e2c105cfd278ab4993562c530a09a53746a65cb680c3923dc6a1c7b55b03dc70"} Oct 13 09:03:40 crc kubenswrapper[4685]: I1013 09:03:40.731288 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3","Type":"ContainerStarted","Data":"2830fa98373c6de60ad60314f67fdf0974dbaf9f674ca8831ad0c2b324e33dfa"} Oct 13 09:03:40 crc kubenswrapper[4685]: I1013 09:03:40.731322 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 13 09:03:40 crc kubenswrapper[4685]: I1013 09:03:40.731332 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.275549 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.372898 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/385f189a-9a2f-42db-a4c1-dd7b332a4f91-log-httpd\") pod \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.373023 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pvpb\" (UniqueName: \"kubernetes.io/projected/385f189a-9a2f-42db-a4c1-dd7b332a4f91-kube-api-access-5pvpb\") pod \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.373088 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-sg-core-conf-yaml\") pod \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.373462 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-config-data\") pod \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.373784 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/385f189a-9a2f-42db-a4c1-dd7b332a4f91-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "385f189a-9a2f-42db-a4c1-dd7b332a4f91" (UID: "385f189a-9a2f-42db-a4c1-dd7b332a4f91"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.373866 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-combined-ca-bundle\") pod \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.373941 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/385f189a-9a2f-42db-a4c1-dd7b332a4f91-run-httpd\") pod \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.373962 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-scripts\") pod \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\" (UID: \"385f189a-9a2f-42db-a4c1-dd7b332a4f91\") " Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.374278 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/385f189a-9a2f-42db-a4c1-dd7b332a4f91-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "385f189a-9a2f-42db-a4c1-dd7b332a4f91" (UID: "385f189a-9a2f-42db-a4c1-dd7b332a4f91"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.374520 4685 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/385f189a-9a2f-42db-a4c1-dd7b332a4f91-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.374539 4685 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/385f189a-9a2f-42db-a4c1-dd7b332a4f91-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.387271 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/385f189a-9a2f-42db-a4c1-dd7b332a4f91-kube-api-access-5pvpb" (OuterVolumeSpecName: "kube-api-access-5pvpb") pod "385f189a-9a2f-42db-a4c1-dd7b332a4f91" (UID: "385f189a-9a2f-42db-a4c1-dd7b332a4f91"). InnerVolumeSpecName "kube-api-access-5pvpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.388080 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-scripts" (OuterVolumeSpecName: "scripts") pod "385f189a-9a2f-42db-a4c1-dd7b332a4f91" (UID: "385f189a-9a2f-42db-a4c1-dd7b332a4f91"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.427457 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "385f189a-9a2f-42db-a4c1-dd7b332a4f91" (UID: "385f189a-9a2f-42db-a4c1-dd7b332a4f91"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.477332 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.478071 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pvpb\" (UniqueName: \"kubernetes.io/projected/385f189a-9a2f-42db-a4c1-dd7b332a4f91-kube-api-access-5pvpb\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.478214 4685 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.478571 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "385f189a-9a2f-42db-a4c1-dd7b332a4f91" (UID: "385f189a-9a2f-42db-a4c1-dd7b332a4f91"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.497003 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-config-data" (OuterVolumeSpecName: "config-data") pod "385f189a-9a2f-42db-a4c1-dd7b332a4f91" (UID: "385f189a-9a2f-42db-a4c1-dd7b332a4f91"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.579600 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.579861 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/385f189a-9a2f-42db-a4c1-dd7b332a4f91-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.740940 4685 generic.go:334] "Generic (PLEG): container finished" podID="385f189a-9a2f-42db-a4c1-dd7b332a4f91" containerID="72ca217c6a84e6ba2a6a56c65d6b450815a3b82fa6b50e64cd15f08f1f71f625" exitCode=0 Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.741099 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"385f189a-9a2f-42db-a4c1-dd7b332a4f91","Type":"ContainerDied","Data":"72ca217c6a84e6ba2a6a56c65d6b450815a3b82fa6b50e64cd15f08f1f71f625"} Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.742025 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"385f189a-9a2f-42db-a4c1-dd7b332a4f91","Type":"ContainerDied","Data":"a4e5db4e63dd41bb0ae291dc2f02d90b9c8ee5238553b86dd2b87c595cc51ff2"} Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.742139 4685 scope.go:117] "RemoveContainer" containerID="0c1c949c42fc6a5194f10fa430e54a5f7fe1b292d7c4a9ea5b54dc5e2f8906e9" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.741157 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.749240 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3","Type":"ContainerStarted","Data":"d24f1abf9f9207417253a11062fa982fef319d820ef7bf7ebcb8e56fd08388f2"} Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.749274 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cf7cb050-76b4-4ee0-b71f-3ad58070fdc3","Type":"ContainerStarted","Data":"0f5846db4a3a062f846ec3b4f7eda22f0c8280d10443fdc037440e91d9951540"} Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.774017 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.775689 4685 scope.go:117] "RemoveContainer" containerID="7f2459f6bdb04ed6437aa0a38c51919274e41eb2c9c2bed408f9f44221a1cf0a" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.797978 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.815507 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:41 crc kubenswrapper[4685]: E1013 09:03:41.816263 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="385f189a-9a2f-42db-a4c1-dd7b332a4f91" containerName="sg-core" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.816357 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="385f189a-9a2f-42db-a4c1-dd7b332a4f91" containerName="sg-core" Oct 13 09:03:41 crc kubenswrapper[4685]: E1013 09:03:41.816429 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="385f189a-9a2f-42db-a4c1-dd7b332a4f91" containerName="ceilometer-notification-agent" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.816509 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="385f189a-9a2f-42db-a4c1-dd7b332a4f91" containerName="ceilometer-notification-agent" Oct 13 09:03:41 crc kubenswrapper[4685]: E1013 09:03:41.816602 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="385f189a-9a2f-42db-a4c1-dd7b332a4f91" containerName="proxy-httpd" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.816677 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="385f189a-9a2f-42db-a4c1-dd7b332a4f91" containerName="proxy-httpd" Oct 13 09:03:41 crc kubenswrapper[4685]: E1013 09:03:41.816756 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="385f189a-9a2f-42db-a4c1-dd7b332a4f91" containerName="ceilometer-central-agent" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.816829 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="385f189a-9a2f-42db-a4c1-dd7b332a4f91" containerName="ceilometer-central-agent" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.817125 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="385f189a-9a2f-42db-a4c1-dd7b332a4f91" containerName="ceilometer-central-agent" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.817227 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="385f189a-9a2f-42db-a4c1-dd7b332a4f91" containerName="proxy-httpd" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.817324 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="385f189a-9a2f-42db-a4c1-dd7b332a4f91" containerName="ceilometer-notification-agent" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.817399 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="385f189a-9a2f-42db-a4c1-dd7b332a4f91" containerName="sg-core" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.819758 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.824880 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.826504 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.826692 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.834221 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.834203172 podStartE2EDuration="3.834203172s" podCreationTimestamp="2025-10-13 09:03:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:03:41.83267108 +0000 UTC m=+1146.980546851" watchObservedRunningTime="2025-10-13 09:03:41.834203172 +0000 UTC m=+1146.982078933" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.874182 4685 scope.go:117] "RemoveContainer" containerID="e2c105cfd278ab4993562c530a09a53746a65cb680c3923dc6a1c7b55b03dc70" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.886173 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " pod="openstack/ceilometer-0" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.887269 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " pod="openstack/ceilometer-0" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.887298 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-run-httpd\") pod \"ceilometer-0\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " pod="openstack/ceilometer-0" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.887362 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-config-data\") pod \"ceilometer-0\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " pod="openstack/ceilometer-0" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.887378 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-log-httpd\") pod \"ceilometer-0\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " pod="openstack/ceilometer-0" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.887402 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-scripts\") pod \"ceilometer-0\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " pod="openstack/ceilometer-0" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.887417 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmlbq\" (UniqueName: \"kubernetes.io/projected/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-kube-api-access-bmlbq\") pod \"ceilometer-0\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " pod="openstack/ceilometer-0" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.906616 4685 scope.go:117] "RemoveContainer" containerID="72ca217c6a84e6ba2a6a56c65d6b450815a3b82fa6b50e64cd15f08f1f71f625" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.932371 4685 scope.go:117] "RemoveContainer" containerID="0c1c949c42fc6a5194f10fa430e54a5f7fe1b292d7c4a9ea5b54dc5e2f8906e9" Oct 13 09:03:41 crc kubenswrapper[4685]: E1013 09:03:41.932861 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c1c949c42fc6a5194f10fa430e54a5f7fe1b292d7c4a9ea5b54dc5e2f8906e9\": container with ID starting with 0c1c949c42fc6a5194f10fa430e54a5f7fe1b292d7c4a9ea5b54dc5e2f8906e9 not found: ID does not exist" containerID="0c1c949c42fc6a5194f10fa430e54a5f7fe1b292d7c4a9ea5b54dc5e2f8906e9" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.932898 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c1c949c42fc6a5194f10fa430e54a5f7fe1b292d7c4a9ea5b54dc5e2f8906e9"} err="failed to get container status \"0c1c949c42fc6a5194f10fa430e54a5f7fe1b292d7c4a9ea5b54dc5e2f8906e9\": rpc error: code = NotFound desc = could not find container \"0c1c949c42fc6a5194f10fa430e54a5f7fe1b292d7c4a9ea5b54dc5e2f8906e9\": container with ID starting with 0c1c949c42fc6a5194f10fa430e54a5f7fe1b292d7c4a9ea5b54dc5e2f8906e9 not found: ID does not exist" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.932931 4685 scope.go:117] "RemoveContainer" containerID="7f2459f6bdb04ed6437aa0a38c51919274e41eb2c9c2bed408f9f44221a1cf0a" Oct 13 09:03:41 crc kubenswrapper[4685]: E1013 09:03:41.933255 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f2459f6bdb04ed6437aa0a38c51919274e41eb2c9c2bed408f9f44221a1cf0a\": container with ID starting with 7f2459f6bdb04ed6437aa0a38c51919274e41eb2c9c2bed408f9f44221a1cf0a not found: ID does not exist" containerID="7f2459f6bdb04ed6437aa0a38c51919274e41eb2c9c2bed408f9f44221a1cf0a" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.933275 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f2459f6bdb04ed6437aa0a38c51919274e41eb2c9c2bed408f9f44221a1cf0a"} err="failed to get container status \"7f2459f6bdb04ed6437aa0a38c51919274e41eb2c9c2bed408f9f44221a1cf0a\": rpc error: code = NotFound desc = could not find container \"7f2459f6bdb04ed6437aa0a38c51919274e41eb2c9c2bed408f9f44221a1cf0a\": container with ID starting with 7f2459f6bdb04ed6437aa0a38c51919274e41eb2c9c2bed408f9f44221a1cf0a not found: ID does not exist" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.933288 4685 scope.go:117] "RemoveContainer" containerID="e2c105cfd278ab4993562c530a09a53746a65cb680c3923dc6a1c7b55b03dc70" Oct 13 09:03:41 crc kubenswrapper[4685]: E1013 09:03:41.933521 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2c105cfd278ab4993562c530a09a53746a65cb680c3923dc6a1c7b55b03dc70\": container with ID starting with e2c105cfd278ab4993562c530a09a53746a65cb680c3923dc6a1c7b55b03dc70 not found: ID does not exist" containerID="e2c105cfd278ab4993562c530a09a53746a65cb680c3923dc6a1c7b55b03dc70" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.933541 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2c105cfd278ab4993562c530a09a53746a65cb680c3923dc6a1c7b55b03dc70"} err="failed to get container status \"e2c105cfd278ab4993562c530a09a53746a65cb680c3923dc6a1c7b55b03dc70\": rpc error: code = NotFound desc = could not find container \"e2c105cfd278ab4993562c530a09a53746a65cb680c3923dc6a1c7b55b03dc70\": container with ID starting with e2c105cfd278ab4993562c530a09a53746a65cb680c3923dc6a1c7b55b03dc70 not found: ID does not exist" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.933553 4685 scope.go:117] "RemoveContainer" containerID="72ca217c6a84e6ba2a6a56c65d6b450815a3b82fa6b50e64cd15f08f1f71f625" Oct 13 09:03:41 crc kubenswrapper[4685]: E1013 09:03:41.933854 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72ca217c6a84e6ba2a6a56c65d6b450815a3b82fa6b50e64cd15f08f1f71f625\": container with ID starting with 72ca217c6a84e6ba2a6a56c65d6b450815a3b82fa6b50e64cd15f08f1f71f625 not found: ID does not exist" containerID="72ca217c6a84e6ba2a6a56c65d6b450815a3b82fa6b50e64cd15f08f1f71f625" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.933889 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72ca217c6a84e6ba2a6a56c65d6b450815a3b82fa6b50e64cd15f08f1f71f625"} err="failed to get container status \"72ca217c6a84e6ba2a6a56c65d6b450815a3b82fa6b50e64cd15f08f1f71f625\": rpc error: code = NotFound desc = could not find container \"72ca217c6a84e6ba2a6a56c65d6b450815a3b82fa6b50e64cd15f08f1f71f625\": container with ID starting with 72ca217c6a84e6ba2a6a56c65d6b450815a3b82fa6b50e64cd15f08f1f71f625 not found: ID does not exist" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.990396 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " pod="openstack/ceilometer-0" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.990450 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-run-httpd\") pod \"ceilometer-0\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " pod="openstack/ceilometer-0" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.990500 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-config-data\") pod \"ceilometer-0\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " pod="openstack/ceilometer-0" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.990524 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-log-httpd\") pod \"ceilometer-0\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " pod="openstack/ceilometer-0" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.990558 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-scripts\") pod \"ceilometer-0\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " pod="openstack/ceilometer-0" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.990579 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmlbq\" (UniqueName: \"kubernetes.io/projected/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-kube-api-access-bmlbq\") pod \"ceilometer-0\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " pod="openstack/ceilometer-0" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.990654 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " pod="openstack/ceilometer-0" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.991460 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-log-httpd\") pod \"ceilometer-0\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " pod="openstack/ceilometer-0" Oct 13 09:03:41 crc kubenswrapper[4685]: I1013 09:03:41.991481 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-run-httpd\") pod \"ceilometer-0\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " pod="openstack/ceilometer-0" Oct 13 09:03:42 crc kubenswrapper[4685]: I1013 09:03:42.003146 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-config-data\") pod \"ceilometer-0\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " pod="openstack/ceilometer-0" Oct 13 09:03:42 crc kubenswrapper[4685]: I1013 09:03:42.005406 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " pod="openstack/ceilometer-0" Oct 13 09:03:42 crc kubenswrapper[4685]: I1013 09:03:42.006108 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-scripts\") pod \"ceilometer-0\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " pod="openstack/ceilometer-0" Oct 13 09:03:42 crc kubenswrapper[4685]: I1013 09:03:42.007487 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmlbq\" (UniqueName: \"kubernetes.io/projected/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-kube-api-access-bmlbq\") pod \"ceilometer-0\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " pod="openstack/ceilometer-0" Oct 13 09:03:42 crc kubenswrapper[4685]: I1013 09:03:42.024896 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " pod="openstack/ceilometer-0" Oct 13 09:03:42 crc kubenswrapper[4685]: I1013 09:03:42.154778 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:03:42 crc kubenswrapper[4685]: I1013 09:03:42.660612 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:42 crc kubenswrapper[4685]: I1013 09:03:42.757631 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2","Type":"ContainerStarted","Data":"c5c87bb40aaffda68d3abe37b7d10e9afd62ccfa18064e0a7223bfe53d339ad3"} Oct 13 09:03:42 crc kubenswrapper[4685]: I1013 09:03:42.758642 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 09:03:42 crc kubenswrapper[4685]: I1013 09:03:42.758852 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 09:03:43 crc kubenswrapper[4685]: I1013 09:03:43.515708 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="385f189a-9a2f-42db-a4c1-dd7b332a4f91" path="/var/lib/kubelet/pods/385f189a-9a2f-42db-a4c1-dd7b332a4f91/volumes" Oct 13 09:03:43 crc kubenswrapper[4685]: I1013 09:03:43.771182 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2","Type":"ContainerStarted","Data":"dbf50d5dcf200187b31c94b576bf01f9b6ceede57593b28f7ec50b23b017d48b"} Oct 13 09:03:43 crc kubenswrapper[4685]: I1013 09:03:43.857617 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 13 09:03:43 crc kubenswrapper[4685]: I1013 09:03:43.857712 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 09:03:43 crc kubenswrapper[4685]: I1013 09:03:43.992736 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 13 09:03:44 crc kubenswrapper[4685]: I1013 09:03:44.788603 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2","Type":"ContainerStarted","Data":"8bc475c0ec7afdda8f09eeec8eabe71eba2443c652da46ac9715a857dc204335"} Oct 13 09:03:45 crc kubenswrapper[4685]: I1013 09:03:45.702763 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:49 crc kubenswrapper[4685]: I1013 09:03:49.235873 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 13 09:03:49 crc kubenswrapper[4685]: I1013 09:03:49.236300 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 13 09:03:49 crc kubenswrapper[4685]: I1013 09:03:49.291191 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 13 09:03:49 crc kubenswrapper[4685]: I1013 09:03:49.295396 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 13 09:03:49 crc kubenswrapper[4685]: I1013 09:03:49.840770 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 13 09:03:49 crc kubenswrapper[4685]: I1013 09:03:49.840837 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 13 09:03:53 crc kubenswrapper[4685]: I1013 09:03:53.876212 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-l4jzf" event={"ID":"11c738eb-e9be-4e64-a38c-da231145e8ab","Type":"ContainerStarted","Data":"d4b0d7d6ad8eb44f0d4a7899d571ce12a286694326a5c8a81ecfa4b0231cabdb"} Oct 13 09:03:53 crc kubenswrapper[4685]: I1013 09:03:53.878071 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2","Type":"ContainerStarted","Data":"17fef511201831a9b582129070f30fa61761b513b57f445ac698bc589d19c0e0"} Oct 13 09:03:53 crc kubenswrapper[4685]: I1013 09:03:53.894284 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-l4jzf" podStartSLOduration=2.119802971 podStartE2EDuration="15.894266533s" podCreationTimestamp="2025-10-13 09:03:38 +0000 UTC" firstStartedPulling="2025-10-13 09:03:39.41755808 +0000 UTC m=+1144.565433841" lastFinishedPulling="2025-10-13 09:03:53.192021632 +0000 UTC m=+1158.339897403" observedRunningTime="2025-10-13 09:03:53.89415639 +0000 UTC m=+1159.042032151" watchObservedRunningTime="2025-10-13 09:03:53.894266533 +0000 UTC m=+1159.042142294" Oct 13 09:03:53 crc kubenswrapper[4685]: I1013 09:03:53.996727 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 13 09:03:53 crc kubenswrapper[4685]: I1013 09:03:53.996824 4685 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 09:03:54 crc kubenswrapper[4685]: I1013 09:03:54.086511 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 13 09:03:55 crc kubenswrapper[4685]: I1013 09:03:55.910995 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2","Type":"ContainerStarted","Data":"60c8fc60fc48e03df8fa772147a9840916f5f4e5358217c030859e2814f268de"} Oct 13 09:03:55 crc kubenswrapper[4685]: I1013 09:03:55.911398 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" containerName="ceilometer-central-agent" containerID="cri-o://dbf50d5dcf200187b31c94b576bf01f9b6ceede57593b28f7ec50b23b017d48b" gracePeriod=30 Oct 13 09:03:55 crc kubenswrapper[4685]: I1013 09:03:55.911735 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 13 09:03:55 crc kubenswrapper[4685]: I1013 09:03:55.912219 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" containerName="sg-core" containerID="cri-o://17fef511201831a9b582129070f30fa61761b513b57f445ac698bc589d19c0e0" gracePeriod=30 Oct 13 09:03:55 crc kubenswrapper[4685]: I1013 09:03:55.912319 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" containerName="proxy-httpd" containerID="cri-o://60c8fc60fc48e03df8fa772147a9840916f5f4e5358217c030859e2814f268de" gracePeriod=30 Oct 13 09:03:55 crc kubenswrapper[4685]: I1013 09:03:55.912387 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" containerName="ceilometer-notification-agent" containerID="cri-o://8bc475c0ec7afdda8f09eeec8eabe71eba2443c652da46ac9715a857dc204335" gracePeriod=30 Oct 13 09:03:55 crc kubenswrapper[4685]: I1013 09:03:55.949069 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.984843731 podStartE2EDuration="14.949041372s" podCreationTimestamp="2025-10-13 09:03:41 +0000 UTC" firstStartedPulling="2025-10-13 09:03:42.667660846 +0000 UTC m=+1147.815536607" lastFinishedPulling="2025-10-13 09:03:54.631858487 +0000 UTC m=+1159.779734248" observedRunningTime="2025-10-13 09:03:55.944412326 +0000 UTC m=+1161.092288127" watchObservedRunningTime="2025-10-13 09:03:55.949041372 +0000 UTC m=+1161.096917133" Oct 13 09:03:56 crc kubenswrapper[4685]: I1013 09:03:56.921225 4685 generic.go:334] "Generic (PLEG): container finished" podID="e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" containerID="60c8fc60fc48e03df8fa772147a9840916f5f4e5358217c030859e2814f268de" exitCode=0 Oct 13 09:03:56 crc kubenswrapper[4685]: I1013 09:03:56.921546 4685 generic.go:334] "Generic (PLEG): container finished" podID="e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" containerID="17fef511201831a9b582129070f30fa61761b513b57f445ac698bc589d19c0e0" exitCode=2 Oct 13 09:03:56 crc kubenswrapper[4685]: I1013 09:03:56.921557 4685 generic.go:334] "Generic (PLEG): container finished" podID="e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" containerID="8bc475c0ec7afdda8f09eeec8eabe71eba2443c652da46ac9715a857dc204335" exitCode=0 Oct 13 09:03:56 crc kubenswrapper[4685]: I1013 09:03:56.921304 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2","Type":"ContainerDied","Data":"60c8fc60fc48e03df8fa772147a9840916f5f4e5358217c030859e2814f268de"} Oct 13 09:03:56 crc kubenswrapper[4685]: I1013 09:03:56.921595 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2","Type":"ContainerDied","Data":"17fef511201831a9b582129070f30fa61761b513b57f445ac698bc589d19c0e0"} Oct 13 09:03:56 crc kubenswrapper[4685]: I1013 09:03:56.921608 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2","Type":"ContainerDied","Data":"8bc475c0ec7afdda8f09eeec8eabe71eba2443c652da46ac9715a857dc204335"} Oct 13 09:03:58 crc kubenswrapper[4685]: I1013 09:03:58.806529 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:03:58 crc kubenswrapper[4685]: I1013 09:03:58.899005 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-log-httpd\") pod \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " Oct 13 09:03:58 crc kubenswrapper[4685]: I1013 09:03:58.899077 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-run-httpd\") pod \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " Oct 13 09:03:58 crc kubenswrapper[4685]: I1013 09:03:58.899122 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-sg-core-conf-yaml\") pod \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " Oct 13 09:03:58 crc kubenswrapper[4685]: I1013 09:03:58.899142 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-combined-ca-bundle\") pod \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " Oct 13 09:03:58 crc kubenswrapper[4685]: I1013 09:03:58.899188 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-config-data\") pod \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " Oct 13 09:03:58 crc kubenswrapper[4685]: I1013 09:03:58.899272 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmlbq\" (UniqueName: \"kubernetes.io/projected/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-kube-api-access-bmlbq\") pod \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " Oct 13 09:03:58 crc kubenswrapper[4685]: I1013 09:03:58.899344 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-scripts\") pod \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " Oct 13 09:03:58 crc kubenswrapper[4685]: I1013 09:03:58.899609 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" (UID: "e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:03:58 crc kubenswrapper[4685]: I1013 09:03:58.899975 4685 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:58 crc kubenswrapper[4685]: I1013 09:03:58.899967 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" (UID: "e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:03:58 crc kubenswrapper[4685]: I1013 09:03:58.905013 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-kube-api-access-bmlbq" (OuterVolumeSpecName: "kube-api-access-bmlbq") pod "e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" (UID: "e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2"). InnerVolumeSpecName "kube-api-access-bmlbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:03:58 crc kubenswrapper[4685]: I1013 09:03:58.910491 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-scripts" (OuterVolumeSpecName: "scripts") pod "e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" (UID: "e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:58 crc kubenswrapper[4685]: I1013 09:03:58.936160 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" (UID: "e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:58 crc kubenswrapper[4685]: I1013 09:03:58.943641 4685 generic.go:334] "Generic (PLEG): container finished" podID="e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" containerID="dbf50d5dcf200187b31c94b576bf01f9b6ceede57593b28f7ec50b23b017d48b" exitCode=0 Oct 13 09:03:58 crc kubenswrapper[4685]: I1013 09:03:58.943695 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2","Type":"ContainerDied","Data":"dbf50d5dcf200187b31c94b576bf01f9b6ceede57593b28f7ec50b23b017d48b"} Oct 13 09:03:58 crc kubenswrapper[4685]: I1013 09:03:58.943719 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:03:58 crc kubenswrapper[4685]: I1013 09:03:58.943735 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2","Type":"ContainerDied","Data":"c5c87bb40aaffda68d3abe37b7d10e9afd62ccfa18064e0a7223bfe53d339ad3"} Oct 13 09:03:58 crc kubenswrapper[4685]: I1013 09:03:58.943759 4685 scope.go:117] "RemoveContainer" containerID="60c8fc60fc48e03df8fa772147a9840916f5f4e5358217c030859e2814f268de" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.004505 4685 scope.go:117] "RemoveContainer" containerID="17fef511201831a9b582129070f30fa61761b513b57f445ac698bc589d19c0e0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.004616 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" (UID: "e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.004753 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-combined-ca-bundle\") pod \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\" (UID: \"e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2\") " Oct 13 09:03:59 crc kubenswrapper[4685]: W1013 09:03:59.004995 4685 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2/volumes/kubernetes.io~secret/combined-ca-bundle Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.005032 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" (UID: "e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.005312 4685 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.005338 4685 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.005375 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.005388 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmlbq\" (UniqueName: \"kubernetes.io/projected/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-kube-api-access-bmlbq\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.005401 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.017338 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-config-data" (OuterVolumeSpecName: "config-data") pod "e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" (UID: "e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.027794 4685 scope.go:117] "RemoveContainer" containerID="8bc475c0ec7afdda8f09eeec8eabe71eba2443c652da46ac9715a857dc204335" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.050612 4685 scope.go:117] "RemoveContainer" containerID="dbf50d5dcf200187b31c94b576bf01f9b6ceede57593b28f7ec50b23b017d48b" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.087298 4685 scope.go:117] "RemoveContainer" containerID="60c8fc60fc48e03df8fa772147a9840916f5f4e5358217c030859e2814f268de" Oct 13 09:03:59 crc kubenswrapper[4685]: E1013 09:03:59.087961 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60c8fc60fc48e03df8fa772147a9840916f5f4e5358217c030859e2814f268de\": container with ID starting with 60c8fc60fc48e03df8fa772147a9840916f5f4e5358217c030859e2814f268de not found: ID does not exist" containerID="60c8fc60fc48e03df8fa772147a9840916f5f4e5358217c030859e2814f268de" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.088017 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60c8fc60fc48e03df8fa772147a9840916f5f4e5358217c030859e2814f268de"} err="failed to get container status \"60c8fc60fc48e03df8fa772147a9840916f5f4e5358217c030859e2814f268de\": rpc error: code = NotFound desc = could not find container \"60c8fc60fc48e03df8fa772147a9840916f5f4e5358217c030859e2814f268de\": container with ID starting with 60c8fc60fc48e03df8fa772147a9840916f5f4e5358217c030859e2814f268de not found: ID does not exist" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.088050 4685 scope.go:117] "RemoveContainer" containerID="17fef511201831a9b582129070f30fa61761b513b57f445ac698bc589d19c0e0" Oct 13 09:03:59 crc kubenswrapper[4685]: E1013 09:03:59.088529 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17fef511201831a9b582129070f30fa61761b513b57f445ac698bc589d19c0e0\": container with ID starting with 17fef511201831a9b582129070f30fa61761b513b57f445ac698bc589d19c0e0 not found: ID does not exist" containerID="17fef511201831a9b582129070f30fa61761b513b57f445ac698bc589d19c0e0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.088549 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17fef511201831a9b582129070f30fa61761b513b57f445ac698bc589d19c0e0"} err="failed to get container status \"17fef511201831a9b582129070f30fa61761b513b57f445ac698bc589d19c0e0\": rpc error: code = NotFound desc = could not find container \"17fef511201831a9b582129070f30fa61761b513b57f445ac698bc589d19c0e0\": container with ID starting with 17fef511201831a9b582129070f30fa61761b513b57f445ac698bc589d19c0e0 not found: ID does not exist" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.088561 4685 scope.go:117] "RemoveContainer" containerID="8bc475c0ec7afdda8f09eeec8eabe71eba2443c652da46ac9715a857dc204335" Oct 13 09:03:59 crc kubenswrapper[4685]: E1013 09:03:59.088884 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bc475c0ec7afdda8f09eeec8eabe71eba2443c652da46ac9715a857dc204335\": container with ID starting with 8bc475c0ec7afdda8f09eeec8eabe71eba2443c652da46ac9715a857dc204335 not found: ID does not exist" containerID="8bc475c0ec7afdda8f09eeec8eabe71eba2443c652da46ac9715a857dc204335" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.088942 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bc475c0ec7afdda8f09eeec8eabe71eba2443c652da46ac9715a857dc204335"} err="failed to get container status \"8bc475c0ec7afdda8f09eeec8eabe71eba2443c652da46ac9715a857dc204335\": rpc error: code = NotFound desc = could not find container \"8bc475c0ec7afdda8f09eeec8eabe71eba2443c652da46ac9715a857dc204335\": container with ID starting with 8bc475c0ec7afdda8f09eeec8eabe71eba2443c652da46ac9715a857dc204335 not found: ID does not exist" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.088973 4685 scope.go:117] "RemoveContainer" containerID="dbf50d5dcf200187b31c94b576bf01f9b6ceede57593b28f7ec50b23b017d48b" Oct 13 09:03:59 crc kubenswrapper[4685]: E1013 09:03:59.089335 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbf50d5dcf200187b31c94b576bf01f9b6ceede57593b28f7ec50b23b017d48b\": container with ID starting with dbf50d5dcf200187b31c94b576bf01f9b6ceede57593b28f7ec50b23b017d48b not found: ID does not exist" containerID="dbf50d5dcf200187b31c94b576bf01f9b6ceede57593b28f7ec50b23b017d48b" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.089360 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbf50d5dcf200187b31c94b576bf01f9b6ceede57593b28f7ec50b23b017d48b"} err="failed to get container status \"dbf50d5dcf200187b31c94b576bf01f9b6ceede57593b28f7ec50b23b017d48b\": rpc error: code = NotFound desc = could not find container \"dbf50d5dcf200187b31c94b576bf01f9b6ceede57593b28f7ec50b23b017d48b\": container with ID starting with dbf50d5dcf200187b31c94b576bf01f9b6ceede57593b28f7ec50b23b017d48b not found: ID does not exist" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.106752 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.285439 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.292705 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.322473 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:59 crc kubenswrapper[4685]: E1013 09:03:59.322809 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" containerName="proxy-httpd" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.322832 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" containerName="proxy-httpd" Oct 13 09:03:59 crc kubenswrapper[4685]: E1013 09:03:59.322847 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" containerName="sg-core" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.322856 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" containerName="sg-core" Oct 13 09:03:59 crc kubenswrapper[4685]: E1013 09:03:59.322866 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" containerName="ceilometer-central-agent" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.322872 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" containerName="ceilometer-central-agent" Oct 13 09:03:59 crc kubenswrapper[4685]: E1013 09:03:59.322885 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" containerName="ceilometer-notification-agent" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.322890 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" containerName="ceilometer-notification-agent" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.323100 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" containerName="ceilometer-notification-agent" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.323112 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" containerName="ceilometer-central-agent" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.323125 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" containerName="proxy-httpd" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.323135 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" containerName="sg-core" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.324631 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.326647 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.335577 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.346207 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.415808 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " pod="openstack/ceilometer-0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.415869 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-log-httpd\") pod \"ceilometer-0\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " pod="openstack/ceilometer-0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.415893 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-config-data\") pod \"ceilometer-0\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " pod="openstack/ceilometer-0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.415935 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-run-httpd\") pod \"ceilometer-0\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " pod="openstack/ceilometer-0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.415973 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " pod="openstack/ceilometer-0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.416046 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsdzh\" (UniqueName: \"kubernetes.io/projected/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-kube-api-access-dsdzh\") pod \"ceilometer-0\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " pod="openstack/ceilometer-0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.416063 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-scripts\") pod \"ceilometer-0\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " pod="openstack/ceilometer-0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.514249 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2" path="/var/lib/kubelet/pods/e2d6e775-9bcd-4d4f-a209-70cde2a7a3a2/volumes" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.517937 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " pod="openstack/ceilometer-0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.517991 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-log-httpd\") pod \"ceilometer-0\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " pod="openstack/ceilometer-0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.518012 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-config-data\") pod \"ceilometer-0\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " pod="openstack/ceilometer-0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.518045 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-run-httpd\") pod \"ceilometer-0\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " pod="openstack/ceilometer-0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.518114 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " pod="openstack/ceilometer-0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.518187 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsdzh\" (UniqueName: \"kubernetes.io/projected/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-kube-api-access-dsdzh\") pod \"ceilometer-0\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " pod="openstack/ceilometer-0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.518203 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-scripts\") pod \"ceilometer-0\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " pod="openstack/ceilometer-0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.519498 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-run-httpd\") pod \"ceilometer-0\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " pod="openstack/ceilometer-0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.521549 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-log-httpd\") pod \"ceilometer-0\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " pod="openstack/ceilometer-0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.523502 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " pod="openstack/ceilometer-0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.524315 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-config-data\") pod \"ceilometer-0\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " pod="openstack/ceilometer-0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.540151 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-scripts\") pod \"ceilometer-0\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " pod="openstack/ceilometer-0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.543359 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " pod="openstack/ceilometer-0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.543708 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsdzh\" (UniqueName: \"kubernetes.io/projected/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-kube-api-access-dsdzh\") pod \"ceilometer-0\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " pod="openstack/ceilometer-0" Oct 13 09:03:59 crc kubenswrapper[4685]: I1013 09:03:59.654467 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:04:00 crc kubenswrapper[4685]: I1013 09:04:00.117636 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:04:00 crc kubenswrapper[4685]: I1013 09:04:00.962867 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103","Type":"ContainerStarted","Data":"5c93a48c8d47d9cce8cca20a1071120c8b4f76312088cda81f5724c286df84f7"} Oct 13 09:04:01 crc kubenswrapper[4685]: I1013 09:04:01.974354 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103","Type":"ContainerStarted","Data":"57934e370879cbfe4d820302b9956f0293dbd0ded9ddd5ff96875099bc630935"} Oct 13 09:04:01 crc kubenswrapper[4685]: I1013 09:04:01.978661 4685 generic.go:334] "Generic (PLEG): container finished" podID="b001d17a-1aea-44ba-86c5-ba6b312156c1" containerID="c44b3a507602babf650e62c8b2c5a5ac1d0f13d11239f82857414365cf156ea7" exitCode=1 Oct 13 09:04:01 crc kubenswrapper[4685]: I1013 09:04:01.978703 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" event={"ID":"b001d17a-1aea-44ba-86c5-ba6b312156c1","Type":"ContainerDied","Data":"c44b3a507602babf650e62c8b2c5a5ac1d0f13d11239f82857414365cf156ea7"} Oct 13 09:04:01 crc kubenswrapper[4685]: I1013 09:04:01.978740 4685 scope.go:117] "RemoveContainer" containerID="1b0c736e43589c77b3b7ea005ec8f55bd09664ac09cc3b4b417ff1bad387941b" Oct 13 09:04:01 crc kubenswrapper[4685]: I1013 09:04:01.979363 4685 scope.go:117] "RemoveContainer" containerID="c44b3a507602babf650e62c8b2c5a5ac1d0f13d11239f82857414365cf156ea7" Oct 13 09:04:01 crc kubenswrapper[4685]: E1013 09:04:01.979574 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:04:02 crc kubenswrapper[4685]: I1013 09:04:02.987829 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103","Type":"ContainerStarted","Data":"6e8bf60e77dda787f80375264cf25c1c9e55e6b418d9a0bfb93bd68d47d9c71d"} Oct 13 09:04:03 crc kubenswrapper[4685]: I1013 09:04:03.442967 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:04:03 crc kubenswrapper[4685]: I1013 09:04:03.443022 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:04:03 crc kubenswrapper[4685]: I1013 09:04:03.443667 4685 scope.go:117] "RemoveContainer" containerID="c44b3a507602babf650e62c8b2c5a5ac1d0f13d11239f82857414365cf156ea7" Oct 13 09:04:03 crc kubenswrapper[4685]: E1013 09:04:03.444042 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:04:04 crc kubenswrapper[4685]: I1013 09:04:04.003640 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103","Type":"ContainerStarted","Data":"5c19ceff41624bd2478d0a35fcb321e096812bbb823fae7adfe826cebd0f2594"} Oct 13 09:04:05 crc kubenswrapper[4685]: I1013 09:04:05.014497 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103","Type":"ContainerStarted","Data":"3d671399559d61833e81b277ebab2cd94594c7316ef9454f073a2cc185dbb6cf"} Oct 13 09:04:05 crc kubenswrapper[4685]: I1013 09:04:05.014841 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 13 09:04:05 crc kubenswrapper[4685]: I1013 09:04:05.044193 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.738500073 podStartE2EDuration="6.044171392s" podCreationTimestamp="2025-10-13 09:03:59 +0000 UTC" firstStartedPulling="2025-10-13 09:04:00.121204284 +0000 UTC m=+1165.269080045" lastFinishedPulling="2025-10-13 09:04:04.426875593 +0000 UTC m=+1169.574751364" observedRunningTime="2025-10-13 09:04:05.03898506 +0000 UTC m=+1170.186860811" watchObservedRunningTime="2025-10-13 09:04:05.044171392 +0000 UTC m=+1170.192047163" Oct 13 09:04:06 crc kubenswrapper[4685]: I1013 09:04:06.029323 4685 generic.go:334] "Generic (PLEG): container finished" podID="11c738eb-e9be-4e64-a38c-da231145e8ab" containerID="d4b0d7d6ad8eb44f0d4a7899d571ce12a286694326a5c8a81ecfa4b0231cabdb" exitCode=0 Oct 13 09:04:06 crc kubenswrapper[4685]: I1013 09:04:06.031010 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-l4jzf" event={"ID":"11c738eb-e9be-4e64-a38c-da231145e8ab","Type":"ContainerDied","Data":"d4b0d7d6ad8eb44f0d4a7899d571ce12a286694326a5c8a81ecfa4b0231cabdb"} Oct 13 09:04:07 crc kubenswrapper[4685]: I1013 09:04:07.394563 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-l4jzf" Oct 13 09:04:07 crc kubenswrapper[4685]: I1013 09:04:07.484059 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjmml\" (UniqueName: \"kubernetes.io/projected/11c738eb-e9be-4e64-a38c-da231145e8ab-kube-api-access-gjmml\") pod \"11c738eb-e9be-4e64-a38c-da231145e8ab\" (UID: \"11c738eb-e9be-4e64-a38c-da231145e8ab\") " Oct 13 09:04:07 crc kubenswrapper[4685]: I1013 09:04:07.484695 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11c738eb-e9be-4e64-a38c-da231145e8ab-config-data\") pod \"11c738eb-e9be-4e64-a38c-da231145e8ab\" (UID: \"11c738eb-e9be-4e64-a38c-da231145e8ab\") " Oct 13 09:04:07 crc kubenswrapper[4685]: I1013 09:04:07.484799 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11c738eb-e9be-4e64-a38c-da231145e8ab-combined-ca-bundle\") pod \"11c738eb-e9be-4e64-a38c-da231145e8ab\" (UID: \"11c738eb-e9be-4e64-a38c-da231145e8ab\") " Oct 13 09:04:07 crc kubenswrapper[4685]: I1013 09:04:07.484938 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11c738eb-e9be-4e64-a38c-da231145e8ab-scripts\") pod \"11c738eb-e9be-4e64-a38c-da231145e8ab\" (UID: \"11c738eb-e9be-4e64-a38c-da231145e8ab\") " Oct 13 09:04:07 crc kubenswrapper[4685]: I1013 09:04:07.491895 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11c738eb-e9be-4e64-a38c-da231145e8ab-kube-api-access-gjmml" (OuterVolumeSpecName: "kube-api-access-gjmml") pod "11c738eb-e9be-4e64-a38c-da231145e8ab" (UID: "11c738eb-e9be-4e64-a38c-da231145e8ab"). InnerVolumeSpecName "kube-api-access-gjmml". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:04:07 crc kubenswrapper[4685]: I1013 09:04:07.492423 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11c738eb-e9be-4e64-a38c-da231145e8ab-scripts" (OuterVolumeSpecName: "scripts") pod "11c738eb-e9be-4e64-a38c-da231145e8ab" (UID: "11c738eb-e9be-4e64-a38c-da231145e8ab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:07 crc kubenswrapper[4685]: I1013 09:04:07.522734 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11c738eb-e9be-4e64-a38c-da231145e8ab-config-data" (OuterVolumeSpecName: "config-data") pod "11c738eb-e9be-4e64-a38c-da231145e8ab" (UID: "11c738eb-e9be-4e64-a38c-da231145e8ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:07 crc kubenswrapper[4685]: I1013 09:04:07.548522 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11c738eb-e9be-4e64-a38c-da231145e8ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11c738eb-e9be-4e64-a38c-da231145e8ab" (UID: "11c738eb-e9be-4e64-a38c-da231145e8ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:07 crc kubenswrapper[4685]: I1013 09:04:07.587713 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11c738eb-e9be-4e64-a38c-da231145e8ab-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:07 crc kubenswrapper[4685]: I1013 09:04:07.587770 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjmml\" (UniqueName: \"kubernetes.io/projected/11c738eb-e9be-4e64-a38c-da231145e8ab-kube-api-access-gjmml\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:07 crc kubenswrapper[4685]: I1013 09:04:07.587790 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11c738eb-e9be-4e64-a38c-da231145e8ab-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:07 crc kubenswrapper[4685]: I1013 09:04:07.587807 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11c738eb-e9be-4e64-a38c-da231145e8ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:08 crc kubenswrapper[4685]: I1013 09:04:08.051572 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-l4jzf" event={"ID":"11c738eb-e9be-4e64-a38c-da231145e8ab","Type":"ContainerDied","Data":"319477a66fdf8aa6e29e69a73736be566414b31696baface8ea779451838061a"} Oct 13 09:04:08 crc kubenswrapper[4685]: I1013 09:04:08.051605 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="319477a66fdf8aa6e29e69a73736be566414b31696baface8ea779451838061a" Oct 13 09:04:08 crc kubenswrapper[4685]: I1013 09:04:08.051632 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-l4jzf" Oct 13 09:04:08 crc kubenswrapper[4685]: I1013 09:04:08.168142 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 13 09:04:08 crc kubenswrapper[4685]: E1013 09:04:08.168792 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11c738eb-e9be-4e64-a38c-da231145e8ab" containerName="nova-cell0-conductor-db-sync" Oct 13 09:04:08 crc kubenswrapper[4685]: I1013 09:04:08.168815 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="11c738eb-e9be-4e64-a38c-da231145e8ab" containerName="nova-cell0-conductor-db-sync" Oct 13 09:04:08 crc kubenswrapper[4685]: I1013 09:04:08.169783 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="11c738eb-e9be-4e64-a38c-da231145e8ab" containerName="nova-cell0-conductor-db-sync" Oct 13 09:04:08 crc kubenswrapper[4685]: I1013 09:04:08.170529 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 13 09:04:08 crc kubenswrapper[4685]: I1013 09:04:08.173266 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 13 09:04:08 crc kubenswrapper[4685]: I1013 09:04:08.173714 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-lqtxf" Oct 13 09:04:08 crc kubenswrapper[4685]: I1013 09:04:08.193301 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 13 09:04:08 crc kubenswrapper[4685]: I1013 09:04:08.302724 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95136327-c720-4660-9e60-adb85bd5afa2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"95136327-c720-4660-9e60-adb85bd5afa2\") " pod="openstack/nova-cell0-conductor-0" Oct 13 09:04:08 crc kubenswrapper[4685]: I1013 09:04:08.303172 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95136327-c720-4660-9e60-adb85bd5afa2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"95136327-c720-4660-9e60-adb85bd5afa2\") " pod="openstack/nova-cell0-conductor-0" Oct 13 09:04:08 crc kubenswrapper[4685]: I1013 09:04:08.303203 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42wt8\" (UniqueName: \"kubernetes.io/projected/95136327-c720-4660-9e60-adb85bd5afa2-kube-api-access-42wt8\") pod \"nova-cell0-conductor-0\" (UID: \"95136327-c720-4660-9e60-adb85bd5afa2\") " pod="openstack/nova-cell0-conductor-0" Oct 13 09:04:08 crc kubenswrapper[4685]: I1013 09:04:08.405280 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95136327-c720-4660-9e60-adb85bd5afa2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"95136327-c720-4660-9e60-adb85bd5afa2\") " pod="openstack/nova-cell0-conductor-0" Oct 13 09:04:08 crc kubenswrapper[4685]: I1013 09:04:08.405449 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95136327-c720-4660-9e60-adb85bd5afa2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"95136327-c720-4660-9e60-adb85bd5afa2\") " pod="openstack/nova-cell0-conductor-0" Oct 13 09:04:08 crc kubenswrapper[4685]: I1013 09:04:08.405490 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42wt8\" (UniqueName: \"kubernetes.io/projected/95136327-c720-4660-9e60-adb85bd5afa2-kube-api-access-42wt8\") pod \"nova-cell0-conductor-0\" (UID: \"95136327-c720-4660-9e60-adb85bd5afa2\") " pod="openstack/nova-cell0-conductor-0" Oct 13 09:04:08 crc kubenswrapper[4685]: I1013 09:04:08.410653 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95136327-c720-4660-9e60-adb85bd5afa2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"95136327-c720-4660-9e60-adb85bd5afa2\") " pod="openstack/nova-cell0-conductor-0" Oct 13 09:04:08 crc kubenswrapper[4685]: I1013 09:04:08.411001 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95136327-c720-4660-9e60-adb85bd5afa2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"95136327-c720-4660-9e60-adb85bd5afa2\") " pod="openstack/nova-cell0-conductor-0" Oct 13 09:04:08 crc kubenswrapper[4685]: I1013 09:04:08.426421 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42wt8\" (UniqueName: \"kubernetes.io/projected/95136327-c720-4660-9e60-adb85bd5afa2-kube-api-access-42wt8\") pod \"nova-cell0-conductor-0\" (UID: \"95136327-c720-4660-9e60-adb85bd5afa2\") " pod="openstack/nova-cell0-conductor-0" Oct 13 09:04:08 crc kubenswrapper[4685]: I1013 09:04:08.502861 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 13 09:04:09 crc kubenswrapper[4685]: W1013 09:04:09.005790 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95136327_c720_4660_9e60_adb85bd5afa2.slice/crio-705a5fe0bb33c42d33d0795c3839221f4127c50731c8f1d6ca37912fd65ae634 WatchSource:0}: Error finding container 705a5fe0bb33c42d33d0795c3839221f4127c50731c8f1d6ca37912fd65ae634: Status 404 returned error can't find the container with id 705a5fe0bb33c42d33d0795c3839221f4127c50731c8f1d6ca37912fd65ae634 Oct 13 09:04:09 crc kubenswrapper[4685]: I1013 09:04:09.020412 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 13 09:04:09 crc kubenswrapper[4685]: I1013 09:04:09.069668 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"95136327-c720-4660-9e60-adb85bd5afa2","Type":"ContainerStarted","Data":"705a5fe0bb33c42d33d0795c3839221f4127c50731c8f1d6ca37912fd65ae634"} Oct 13 09:04:10 crc kubenswrapper[4685]: I1013 09:04:10.080801 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"95136327-c720-4660-9e60-adb85bd5afa2","Type":"ContainerStarted","Data":"543f70491a7f880b3ad00d88c3a65c5954f4b31e29ca4ca9b7db7a2d965611e1"} Oct 13 09:04:10 crc kubenswrapper[4685]: I1013 09:04:10.082258 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 13 09:04:10 crc kubenswrapper[4685]: I1013 09:04:10.101293 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.10126774 podStartE2EDuration="2.10126774s" podCreationTimestamp="2025-10-13 09:04:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:04:10.099412529 +0000 UTC m=+1175.247288290" watchObservedRunningTime="2025-10-13 09:04:10.10126774 +0000 UTC m=+1175.249143521" Oct 13 09:04:18 crc kubenswrapper[4685]: I1013 09:04:18.503678 4685 scope.go:117] "RemoveContainer" containerID="c44b3a507602babf650e62c8b2c5a5ac1d0f13d11239f82857414365cf156ea7" Oct 13 09:04:18 crc kubenswrapper[4685]: I1013 09:04:18.543321 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 13 09:04:18 crc kubenswrapper[4685]: I1013 09:04:18.999052 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-5psdz"] Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.000637 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5psdz" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.005533 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.013166 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.018209 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-5psdz"] Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.154730 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-config-data\") pod \"nova-cell0-cell-mapping-5psdz\" (UID: \"5c0a4dc6-640e-4aea-9865-aadb943fcc6a\") " pod="openstack/nova-cell0-cell-mapping-5psdz" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.154772 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-scripts\") pod \"nova-cell0-cell-mapping-5psdz\" (UID: \"5c0a4dc6-640e-4aea-9865-aadb943fcc6a\") " pod="openstack/nova-cell0-cell-mapping-5psdz" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.154822 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-5psdz\" (UID: \"5c0a4dc6-640e-4aea-9865-aadb943fcc6a\") " pod="openstack/nova-cell0-cell-mapping-5psdz" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.154843 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvjdx\" (UniqueName: \"kubernetes.io/projected/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-kube-api-access-dvjdx\") pod \"nova-cell0-cell-mapping-5psdz\" (UID: \"5c0a4dc6-640e-4aea-9865-aadb943fcc6a\") " pod="openstack/nova-cell0-cell-mapping-5psdz" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.196725 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" event={"ID":"b001d17a-1aea-44ba-86c5-ba6b312156c1","Type":"ContainerStarted","Data":"6f8ac1889cdb78c30c1167f41924fc8421f395f14a8be5ba49377fbd32fe2472"} Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.197054 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.242080 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.243566 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.256347 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-config-data\") pod \"nova-cell0-cell-mapping-5psdz\" (UID: \"5c0a4dc6-640e-4aea-9865-aadb943fcc6a\") " pod="openstack/nova-cell0-cell-mapping-5psdz" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.256387 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-scripts\") pod \"nova-cell0-cell-mapping-5psdz\" (UID: \"5c0a4dc6-640e-4aea-9865-aadb943fcc6a\") " pod="openstack/nova-cell0-cell-mapping-5psdz" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.256439 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-5psdz\" (UID: \"5c0a4dc6-640e-4aea-9865-aadb943fcc6a\") " pod="openstack/nova-cell0-cell-mapping-5psdz" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.256459 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvjdx\" (UniqueName: \"kubernetes.io/projected/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-kube-api-access-dvjdx\") pod \"nova-cell0-cell-mapping-5psdz\" (UID: \"5c0a4dc6-640e-4aea-9865-aadb943fcc6a\") " pod="openstack/nova-cell0-cell-mapping-5psdz" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.256877 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.270157 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-5psdz\" (UID: \"5c0a4dc6-640e-4aea-9865-aadb943fcc6a\") " pod="openstack/nova-cell0-cell-mapping-5psdz" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.271414 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-scripts\") pod \"nova-cell0-cell-mapping-5psdz\" (UID: \"5c0a4dc6-640e-4aea-9865-aadb943fcc6a\") " pod="openstack/nova-cell0-cell-mapping-5psdz" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.286668 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-config-data\") pod \"nova-cell0-cell-mapping-5psdz\" (UID: \"5c0a4dc6-640e-4aea-9865-aadb943fcc6a\") " pod="openstack/nova-cell0-cell-mapping-5psdz" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.294850 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.296363 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.303779 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.319999 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.332936 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvjdx\" (UniqueName: \"kubernetes.io/projected/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-kube-api-access-dvjdx\") pod \"nova-cell0-cell-mapping-5psdz\" (UID: \"5c0a4dc6-640e-4aea-9865-aadb943fcc6a\") " pod="openstack/nova-cell0-cell-mapping-5psdz" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.349099 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.364953 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2n9x\" (UniqueName: \"kubernetes.io/projected/e50b5342-8da0-4176-b977-84860f396352-kube-api-access-z2n9x\") pod \"nova-metadata-0\" (UID: \"e50b5342-8da0-4176-b977-84860f396352\") " pod="openstack/nova-metadata-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.365045 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e50b5342-8da0-4176-b977-84860f396352-config-data\") pod \"nova-metadata-0\" (UID: \"e50b5342-8da0-4176-b977-84860f396352\") " pod="openstack/nova-metadata-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.365113 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e50b5342-8da0-4176-b977-84860f396352-logs\") pod \"nova-metadata-0\" (UID: \"e50b5342-8da0-4176-b977-84860f396352\") " pod="openstack/nova-metadata-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.365177 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50b5342-8da0-4176-b977-84860f396352-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e50b5342-8da0-4176-b977-84860f396352\") " pod="openstack/nova-metadata-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.433529 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.435064 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.443297 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.470644 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b00bb156-f7b2-485b-8dce-424707695f9b-logs\") pod \"nova-api-0\" (UID: \"b00bb156-f7b2-485b-8dce-424707695f9b\") " pod="openstack/nova-api-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.470697 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e50b5342-8da0-4176-b977-84860f396352-logs\") pod \"nova-metadata-0\" (UID: \"e50b5342-8da0-4176-b977-84860f396352\") " pod="openstack/nova-metadata-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.470798 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50b5342-8da0-4176-b977-84860f396352-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e50b5342-8da0-4176-b977-84860f396352\") " pod="openstack/nova-metadata-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.470845 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b00bb156-f7b2-485b-8dce-424707695f9b-config-data\") pod \"nova-api-0\" (UID: \"b00bb156-f7b2-485b-8dce-424707695f9b\") " pod="openstack/nova-api-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.470880 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6wtw\" (UniqueName: \"kubernetes.io/projected/b00bb156-f7b2-485b-8dce-424707695f9b-kube-api-access-x6wtw\") pod \"nova-api-0\" (UID: \"b00bb156-f7b2-485b-8dce-424707695f9b\") " pod="openstack/nova-api-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.470926 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2n9x\" (UniqueName: \"kubernetes.io/projected/e50b5342-8da0-4176-b977-84860f396352-kube-api-access-z2n9x\") pod \"nova-metadata-0\" (UID: \"e50b5342-8da0-4176-b977-84860f396352\") " pod="openstack/nova-metadata-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.471023 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e50b5342-8da0-4176-b977-84860f396352-config-data\") pod \"nova-metadata-0\" (UID: \"e50b5342-8da0-4176-b977-84860f396352\") " pod="openstack/nova-metadata-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.471059 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b00bb156-f7b2-485b-8dce-424707695f9b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b00bb156-f7b2-485b-8dce-424707695f9b\") " pod="openstack/nova-api-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.482670 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.483776 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.486477 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e50b5342-8da0-4176-b977-84860f396352-logs\") pod \"nova-metadata-0\" (UID: \"e50b5342-8da0-4176-b977-84860f396352\") " pod="openstack/nova-metadata-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.494423 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.503268 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-nv9ck"] Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.506533 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e50b5342-8da0-4176-b977-84860f396352-config-data\") pod \"nova-metadata-0\" (UID: \"e50b5342-8da0-4176-b977-84860f396352\") " pod="openstack/nova-metadata-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.507730 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.534527 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50b5342-8da0-4176-b977-84860f396352-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e50b5342-8da0-4176-b977-84860f396352\") " pod="openstack/nova-metadata-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.540247 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2n9x\" (UniqueName: \"kubernetes.io/projected/e50b5342-8da0-4176-b977-84860f396352-kube-api-access-z2n9x\") pod \"nova-metadata-0\" (UID: \"e50b5342-8da0-4176-b977-84860f396352\") " pod="openstack/nova-metadata-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.572543 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b00bb156-f7b2-485b-8dce-424707695f9b-logs\") pod \"nova-api-0\" (UID: \"b00bb156-f7b2-485b-8dce-424707695f9b\") " pod="openstack/nova-api-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.572802 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd3dba3c-5b8e-45a5-a38a-528eea861a6a-config-data\") pod \"nova-scheduler-0\" (UID: \"bd3dba3c-5b8e-45a5-a38a-528eea861a6a\") " pod="openstack/nova-scheduler-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.572935 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b00bb156-f7b2-485b-8dce-424707695f9b-config-data\") pod \"nova-api-0\" (UID: \"b00bb156-f7b2-485b-8dce-424707695f9b\") " pod="openstack/nova-api-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.573030 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd3dba3c-5b8e-45a5-a38a-528eea861a6a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bd3dba3c-5b8e-45a5-a38a-528eea861a6a\") " pod="openstack/nova-scheduler-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.573112 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6wtw\" (UniqueName: \"kubernetes.io/projected/b00bb156-f7b2-485b-8dce-424707695f9b-kube-api-access-x6wtw\") pod \"nova-api-0\" (UID: \"b00bb156-f7b2-485b-8dce-424707695f9b\") " pod="openstack/nova-api-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.573199 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8njl\" (UniqueName: \"kubernetes.io/projected/bd3dba3c-5b8e-45a5-a38a-528eea861a6a-kube-api-access-d8njl\") pod \"nova-scheduler-0\" (UID: \"bd3dba3c-5b8e-45a5-a38a-528eea861a6a\") " pod="openstack/nova-scheduler-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.573391 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b00bb156-f7b2-485b-8dce-424707695f9b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b00bb156-f7b2-485b-8dce-424707695f9b\") " pod="openstack/nova-api-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.573079 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b00bb156-f7b2-485b-8dce-424707695f9b-logs\") pod \"nova-api-0\" (UID: \"b00bb156-f7b2-485b-8dce-424707695f9b\") " pod="openstack/nova-api-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.594222 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b00bb156-f7b2-485b-8dce-424707695f9b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b00bb156-f7b2-485b-8dce-424707695f9b\") " pod="openstack/nova-api-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.601269 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6wtw\" (UniqueName: \"kubernetes.io/projected/b00bb156-f7b2-485b-8dce-424707695f9b-kube-api-access-x6wtw\") pod \"nova-api-0\" (UID: \"b00bb156-f7b2-485b-8dce-424707695f9b\") " pod="openstack/nova-api-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.604156 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.617735 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5psdz" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.625694 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b00bb156-f7b2-485b-8dce-424707695f9b-config-data\") pod \"nova-api-0\" (UID: \"b00bb156-f7b2-485b-8dce-424707695f9b\") " pod="openstack/nova-api-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.675136 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd3dba3c-5b8e-45a5-a38a-528eea861a6a-config-data\") pod \"nova-scheduler-0\" (UID: \"bd3dba3c-5b8e-45a5-a38a-528eea861a6a\") " pod="openstack/nova-scheduler-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.675871 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-nv9ck\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.675926 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0177ea10-0a27-4cfb-99d8-0454b56e1863-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0177ea10-0a27-4cfb-99d8-0454b56e1863\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.675962 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd3dba3c-5b8e-45a5-a38a-528eea861a6a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bd3dba3c-5b8e-45a5-a38a-528eea861a6a\") " pod="openstack/nova-scheduler-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.675986 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbvff\" (UniqueName: \"kubernetes.io/projected/0177ea10-0a27-4cfb-99d8-0454b56e1863-kube-api-access-cbvff\") pod \"nova-cell1-novncproxy-0\" (UID: \"0177ea10-0a27-4cfb-99d8-0454b56e1863\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.676005 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8njl\" (UniqueName: \"kubernetes.io/projected/bd3dba3c-5b8e-45a5-a38a-528eea861a6a-kube-api-access-d8njl\") pod \"nova-scheduler-0\" (UID: \"bd3dba3c-5b8e-45a5-a38a-528eea861a6a\") " pod="openstack/nova-scheduler-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.676028 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-dns-svc\") pod \"dnsmasq-dns-bccf8f775-nv9ck\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.676052 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-config\") pod \"dnsmasq-dns-bccf8f775-nv9ck\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.676084 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0177ea10-0a27-4cfb-99d8-0454b56e1863-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0177ea10-0a27-4cfb-99d8-0454b56e1863\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.676122 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwvs8\" (UniqueName: \"kubernetes.io/projected/7f5ec6ac-31a4-4d32-a600-2c274d9535df-kube-api-access-dwvs8\") pod \"dnsmasq-dns-bccf8f775-nv9ck\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.676151 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-nv9ck\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.676179 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-nv9ck\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.692218 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd3dba3c-5b8e-45a5-a38a-528eea861a6a-config-data\") pod \"nova-scheduler-0\" (UID: \"bd3dba3c-5b8e-45a5-a38a-528eea861a6a\") " pod="openstack/nova-scheduler-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.692509 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-nv9ck"] Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.717398 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd3dba3c-5b8e-45a5-a38a-528eea861a6a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bd3dba3c-5b8e-45a5-a38a-528eea861a6a\") " pod="openstack/nova-scheduler-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.726952 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8njl\" (UniqueName: \"kubernetes.io/projected/bd3dba3c-5b8e-45a5-a38a-528eea861a6a-kube-api-access-d8njl\") pod \"nova-scheduler-0\" (UID: \"bd3dba3c-5b8e-45a5-a38a-528eea861a6a\") " pod="openstack/nova-scheduler-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.731941 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.744884 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.778147 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-nv9ck\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.778297 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-nv9ck\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.778448 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-nv9ck\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.778483 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0177ea10-0a27-4cfb-99d8-0454b56e1863-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0177ea10-0a27-4cfb-99d8-0454b56e1863\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.778530 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbvff\" (UniqueName: \"kubernetes.io/projected/0177ea10-0a27-4cfb-99d8-0454b56e1863-kube-api-access-cbvff\") pod \"nova-cell1-novncproxy-0\" (UID: \"0177ea10-0a27-4cfb-99d8-0454b56e1863\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.778558 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-dns-svc\") pod \"dnsmasq-dns-bccf8f775-nv9ck\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.778581 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-config\") pod \"dnsmasq-dns-bccf8f775-nv9ck\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.778623 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0177ea10-0a27-4cfb-99d8-0454b56e1863-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0177ea10-0a27-4cfb-99d8-0454b56e1863\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.778654 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwvs8\" (UniqueName: \"kubernetes.io/projected/7f5ec6ac-31a4-4d32-a600-2c274d9535df-kube-api-access-dwvs8\") pod \"dnsmasq-dns-bccf8f775-nv9ck\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.780185 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-dns-svc\") pod \"dnsmasq-dns-bccf8f775-nv9ck\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.780990 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-nv9ck\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.781145 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-nv9ck\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.781403 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-config\") pod \"dnsmasq-dns-bccf8f775-nv9ck\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.784401 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-nv9ck\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.785580 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0177ea10-0a27-4cfb-99d8-0454b56e1863-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0177ea10-0a27-4cfb-99d8-0454b56e1863\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.794878 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0177ea10-0a27-4cfb-99d8-0454b56e1863-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0177ea10-0a27-4cfb-99d8-0454b56e1863\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.799251 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwvs8\" (UniqueName: \"kubernetes.io/projected/7f5ec6ac-31a4-4d32-a600-2c274d9535df-kube-api-access-dwvs8\") pod \"dnsmasq-dns-bccf8f775-nv9ck\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.808681 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbvff\" (UniqueName: \"kubernetes.io/projected/0177ea10-0a27-4cfb-99d8-0454b56e1863-kube-api-access-cbvff\") pod \"nova-cell1-novncproxy-0\" (UID: \"0177ea10-0a27-4cfb-99d8-0454b56e1863\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.814903 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.831287 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.852887 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:04:19 crc kubenswrapper[4685]: I1013 09:04:19.873505 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 09:04:20 crc kubenswrapper[4685]: I1013 09:04:20.349737 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-5psdz"] Oct 13 09:04:20 crc kubenswrapper[4685]: W1013 09:04:20.351684 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c0a4dc6_640e_4aea_9865_aadb943fcc6a.slice/crio-addc0933f45f0d05018e86dc9981484f4f96fcf6d031da30c2877a7d92e76c69 WatchSource:0}: Error finding container addc0933f45f0d05018e86dc9981484f4f96fcf6d031da30c2877a7d92e76c69: Status 404 returned error can't find the container with id addc0933f45f0d05018e86dc9981484f4f96fcf6d031da30c2877a7d92e76c69 Oct 13 09:04:20 crc kubenswrapper[4685]: I1013 09:04:20.413784 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 09:04:20 crc kubenswrapper[4685]: I1013 09:04:20.500449 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 09:04:20 crc kubenswrapper[4685]: I1013 09:04:20.753569 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 13 09:04:20 crc kubenswrapper[4685]: I1013 09:04:20.795094 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-nv9ck"] Oct 13 09:04:20 crc kubenswrapper[4685]: I1013 09:04:20.877723 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 13 09:04:20 crc kubenswrapper[4685]: I1013 09:04:20.892369 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-wvrpr"] Oct 13 09:04:20 crc kubenswrapper[4685]: I1013 09:04:20.893769 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-wvrpr" Oct 13 09:04:20 crc kubenswrapper[4685]: I1013 09:04:20.902238 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 13 09:04:20 crc kubenswrapper[4685]: I1013 09:04:20.902386 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 13 09:04:20 crc kubenswrapper[4685]: I1013 09:04:20.911207 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-wvrpr"] Oct 13 09:04:20 crc kubenswrapper[4685]: I1013 09:04:20.999521 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mshkh\" (UniqueName: \"kubernetes.io/projected/b8d4a69c-f774-44b5-bb13-5340f07027be-kube-api-access-mshkh\") pod \"nova-cell1-conductor-db-sync-wvrpr\" (UID: \"b8d4a69c-f774-44b5-bb13-5340f07027be\") " pod="openstack/nova-cell1-conductor-db-sync-wvrpr" Oct 13 09:04:20 crc kubenswrapper[4685]: I1013 09:04:20.999897 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8d4a69c-f774-44b5-bb13-5340f07027be-config-data\") pod \"nova-cell1-conductor-db-sync-wvrpr\" (UID: \"b8d4a69c-f774-44b5-bb13-5340f07027be\") " pod="openstack/nova-cell1-conductor-db-sync-wvrpr" Oct 13 09:04:21 crc kubenswrapper[4685]: I1013 09:04:21.000395 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8d4a69c-f774-44b5-bb13-5340f07027be-scripts\") pod \"nova-cell1-conductor-db-sync-wvrpr\" (UID: \"b8d4a69c-f774-44b5-bb13-5340f07027be\") " pod="openstack/nova-cell1-conductor-db-sync-wvrpr" Oct 13 09:04:21 crc kubenswrapper[4685]: I1013 09:04:21.000754 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8d4a69c-f774-44b5-bb13-5340f07027be-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-wvrpr\" (UID: \"b8d4a69c-f774-44b5-bb13-5340f07027be\") " pod="openstack/nova-cell1-conductor-db-sync-wvrpr" Oct 13 09:04:21 crc kubenswrapper[4685]: I1013 09:04:21.103043 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8d4a69c-f774-44b5-bb13-5340f07027be-config-data\") pod \"nova-cell1-conductor-db-sync-wvrpr\" (UID: \"b8d4a69c-f774-44b5-bb13-5340f07027be\") " pod="openstack/nova-cell1-conductor-db-sync-wvrpr" Oct 13 09:04:21 crc kubenswrapper[4685]: I1013 09:04:21.104787 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8d4a69c-f774-44b5-bb13-5340f07027be-scripts\") pod \"nova-cell1-conductor-db-sync-wvrpr\" (UID: \"b8d4a69c-f774-44b5-bb13-5340f07027be\") " pod="openstack/nova-cell1-conductor-db-sync-wvrpr" Oct 13 09:04:21 crc kubenswrapper[4685]: I1013 09:04:21.105110 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8d4a69c-f774-44b5-bb13-5340f07027be-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-wvrpr\" (UID: \"b8d4a69c-f774-44b5-bb13-5340f07027be\") " pod="openstack/nova-cell1-conductor-db-sync-wvrpr" Oct 13 09:04:21 crc kubenswrapper[4685]: I1013 09:04:21.105401 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mshkh\" (UniqueName: \"kubernetes.io/projected/b8d4a69c-f774-44b5-bb13-5340f07027be-kube-api-access-mshkh\") pod \"nova-cell1-conductor-db-sync-wvrpr\" (UID: \"b8d4a69c-f774-44b5-bb13-5340f07027be\") " pod="openstack/nova-cell1-conductor-db-sync-wvrpr" Oct 13 09:04:21 crc kubenswrapper[4685]: I1013 09:04:21.108368 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8d4a69c-f774-44b5-bb13-5340f07027be-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-wvrpr\" (UID: \"b8d4a69c-f774-44b5-bb13-5340f07027be\") " pod="openstack/nova-cell1-conductor-db-sync-wvrpr" Oct 13 09:04:21 crc kubenswrapper[4685]: I1013 09:04:21.109605 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8d4a69c-f774-44b5-bb13-5340f07027be-scripts\") pod \"nova-cell1-conductor-db-sync-wvrpr\" (UID: \"b8d4a69c-f774-44b5-bb13-5340f07027be\") " pod="openstack/nova-cell1-conductor-db-sync-wvrpr" Oct 13 09:04:21 crc kubenswrapper[4685]: I1013 09:04:21.109882 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8d4a69c-f774-44b5-bb13-5340f07027be-config-data\") pod \"nova-cell1-conductor-db-sync-wvrpr\" (UID: \"b8d4a69c-f774-44b5-bb13-5340f07027be\") " pod="openstack/nova-cell1-conductor-db-sync-wvrpr" Oct 13 09:04:21 crc kubenswrapper[4685]: I1013 09:04:21.120855 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mshkh\" (UniqueName: \"kubernetes.io/projected/b8d4a69c-f774-44b5-bb13-5340f07027be-kube-api-access-mshkh\") pod \"nova-cell1-conductor-db-sync-wvrpr\" (UID: \"b8d4a69c-f774-44b5-bb13-5340f07027be\") " pod="openstack/nova-cell1-conductor-db-sync-wvrpr" Oct 13 09:04:21 crc kubenswrapper[4685]: I1013 09:04:21.221739 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-wvrpr" Oct 13 09:04:21 crc kubenswrapper[4685]: I1013 09:04:21.233774 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e50b5342-8da0-4176-b977-84860f396352","Type":"ContainerStarted","Data":"4f0ee6be0519670699bd39a3484ffe05aaa96d4e990ad18c0a04b1fadbea3f44"} Oct 13 09:04:21 crc kubenswrapper[4685]: I1013 09:04:21.235789 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b00bb156-f7b2-485b-8dce-424707695f9b","Type":"ContainerStarted","Data":"0c4cb5e6cf0dcadeea5dd664332c7af4c9866420f5655e81290e3346861b8741"} Oct 13 09:04:21 crc kubenswrapper[4685]: I1013 09:04:21.238000 4685 generic.go:334] "Generic (PLEG): container finished" podID="7f5ec6ac-31a4-4d32-a600-2c274d9535df" containerID="dfa6ba7e634681fda898f0740e271c94b43bc2404d62bfe268cd1ffee6f95b24" exitCode=0 Oct 13 09:04:21 crc kubenswrapper[4685]: I1013 09:04:21.238071 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" event={"ID":"7f5ec6ac-31a4-4d32-a600-2c274d9535df","Type":"ContainerDied","Data":"dfa6ba7e634681fda898f0740e271c94b43bc2404d62bfe268cd1ffee6f95b24"} Oct 13 09:04:21 crc kubenswrapper[4685]: I1013 09:04:21.238098 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" event={"ID":"7f5ec6ac-31a4-4d32-a600-2c274d9535df","Type":"ContainerStarted","Data":"2414585a93b2c26acd6f05cc17e9abbf003b285207cc1236089556089138fd27"} Oct 13 09:04:21 crc kubenswrapper[4685]: I1013 09:04:21.239335 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0177ea10-0a27-4cfb-99d8-0454b56e1863","Type":"ContainerStarted","Data":"0a03599f7c4c422e3749e32157e261e205839bbdeaf83bd7daf38c74bd58dc10"} Oct 13 09:04:21 crc kubenswrapper[4685]: I1013 09:04:21.241972 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bd3dba3c-5b8e-45a5-a38a-528eea861a6a","Type":"ContainerStarted","Data":"62e5efb29fbff7231034cd604144685597d1735bf88ef53cd49f0cd82896dc0f"} Oct 13 09:04:21 crc kubenswrapper[4685]: I1013 09:04:21.245622 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5psdz" event={"ID":"5c0a4dc6-640e-4aea-9865-aadb943fcc6a","Type":"ContainerStarted","Data":"cbef8ef43372806b6c937531fa6dc0625fe37f464ed56e90a130285aadb2a357"} Oct 13 09:04:21 crc kubenswrapper[4685]: I1013 09:04:21.245667 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5psdz" event={"ID":"5c0a4dc6-640e-4aea-9865-aadb943fcc6a","Type":"ContainerStarted","Data":"addc0933f45f0d05018e86dc9981484f4f96fcf6d031da30c2877a7d92e76c69"} Oct 13 09:04:21 crc kubenswrapper[4685]: I1013 09:04:21.286815 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-5psdz" podStartSLOduration=3.286797799 podStartE2EDuration="3.286797799s" podCreationTimestamp="2025-10-13 09:04:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:04:21.283334025 +0000 UTC m=+1186.431209786" watchObservedRunningTime="2025-10-13 09:04:21.286797799 +0000 UTC m=+1186.434673560" Oct 13 09:04:21 crc kubenswrapper[4685]: I1013 09:04:21.862741 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-wvrpr"] Oct 13 09:04:21 crc kubenswrapper[4685]: W1013 09:04:21.916702 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8d4a69c_f774_44b5_bb13_5340f07027be.slice/crio-042aa059d6446d3dfc9f2ba5bf0098b36710242a80840e8e1c241714f46be975 WatchSource:0}: Error finding container 042aa059d6446d3dfc9f2ba5bf0098b36710242a80840e8e1c241714f46be975: Status 404 returned error can't find the container with id 042aa059d6446d3dfc9f2ba5bf0098b36710242a80840e8e1c241714f46be975 Oct 13 09:04:22 crc kubenswrapper[4685]: I1013 09:04:22.299684 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" event={"ID":"7f5ec6ac-31a4-4d32-a600-2c274d9535df","Type":"ContainerStarted","Data":"7ce9dadc75ad7f16a641fb2b4705e647595b8b2a879975af4bb44706eca7143e"} Oct 13 09:04:22 crc kubenswrapper[4685]: I1013 09:04:22.300030 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:04:22 crc kubenswrapper[4685]: I1013 09:04:22.306332 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-wvrpr" event={"ID":"b8d4a69c-f774-44b5-bb13-5340f07027be","Type":"ContainerStarted","Data":"a110131aaa1c74e0bcd8e70b66e2ec1b6e514ab7428b13aeff24b729d18c411c"} Oct 13 09:04:22 crc kubenswrapper[4685]: I1013 09:04:22.306377 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-wvrpr" event={"ID":"b8d4a69c-f774-44b5-bb13-5340f07027be","Type":"ContainerStarted","Data":"042aa059d6446d3dfc9f2ba5bf0098b36710242a80840e8e1c241714f46be975"} Oct 13 09:04:22 crc kubenswrapper[4685]: I1013 09:04:22.363462 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-wvrpr" podStartSLOduration=2.363442935 podStartE2EDuration="2.363442935s" podCreationTimestamp="2025-10-13 09:04:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:04:22.358577592 +0000 UTC m=+1187.506453353" watchObservedRunningTime="2025-10-13 09:04:22.363442935 +0000 UTC m=+1187.511318706" Oct 13 09:04:22 crc kubenswrapper[4685]: I1013 09:04:22.367455 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" podStartSLOduration=3.367438164 podStartE2EDuration="3.367438164s" podCreationTimestamp="2025-10-13 09:04:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:04:22.335631807 +0000 UTC m=+1187.483507568" watchObservedRunningTime="2025-10-13 09:04:22.367438164 +0000 UTC m=+1187.515313945" Oct 13 09:04:23 crc kubenswrapper[4685]: I1013 09:04:23.431123 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 09:04:23 crc kubenswrapper[4685]: I1013 09:04:23.455490 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:04:23 crc kubenswrapper[4685]: I1013 09:04:23.461477 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 13 09:04:26 crc kubenswrapper[4685]: I1013 09:04:26.355202 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0177ea10-0a27-4cfb-99d8-0454b56e1863","Type":"ContainerStarted","Data":"6a2310a77e69619d81f679dd96fae0d10c00fa331514a47bc1fdc54a76fd5665"} Oct 13 09:04:26 crc kubenswrapper[4685]: I1013 09:04:26.358429 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="0177ea10-0a27-4cfb-99d8-0454b56e1863" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://6a2310a77e69619d81f679dd96fae0d10c00fa331514a47bc1fdc54a76fd5665" gracePeriod=30 Oct 13 09:04:26 crc kubenswrapper[4685]: I1013 09:04:26.361460 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bd3dba3c-5b8e-45a5-a38a-528eea861a6a","Type":"ContainerStarted","Data":"432d2d8f755a059bf7bb2476e30c65637170774997900033d20f812b8a076a77"} Oct 13 09:04:26 crc kubenswrapper[4685]: I1013 09:04:26.365088 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e50b5342-8da0-4176-b977-84860f396352","Type":"ContainerStarted","Data":"be83a0a70a203abfdf60f9f72920ffb147a6d6e328cff0c6b5ac8f2148750d20"} Oct 13 09:04:26 crc kubenswrapper[4685]: I1013 09:04:26.365126 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e50b5342-8da0-4176-b977-84860f396352","Type":"ContainerStarted","Data":"6a833e7e95003fca3f99dfc98ff560809af66bf9f2337ef23aeed7abcdebdd22"} Oct 13 09:04:26 crc kubenswrapper[4685]: I1013 09:04:26.365225 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e50b5342-8da0-4176-b977-84860f396352" containerName="nova-metadata-log" containerID="cri-o://6a833e7e95003fca3f99dfc98ff560809af66bf9f2337ef23aeed7abcdebdd22" gracePeriod=30 Oct 13 09:04:26 crc kubenswrapper[4685]: I1013 09:04:26.365463 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e50b5342-8da0-4176-b977-84860f396352" containerName="nova-metadata-metadata" containerID="cri-o://be83a0a70a203abfdf60f9f72920ffb147a6d6e328cff0c6b5ac8f2148750d20" gracePeriod=30 Oct 13 09:04:26 crc kubenswrapper[4685]: I1013 09:04:26.370594 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b00bb156-f7b2-485b-8dce-424707695f9b","Type":"ContainerStarted","Data":"01f99c9948e44ee05ef7ac3ee3c6de6b135c46be801bf098c2cdd88e13b881b2"} Oct 13 09:04:26 crc kubenswrapper[4685]: I1013 09:04:26.370641 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b00bb156-f7b2-485b-8dce-424707695f9b","Type":"ContainerStarted","Data":"f5ddbd46160fd2af94ec224e443d414eb639f90b33562a8997c3bc3dc0103af8"} Oct 13 09:04:26 crc kubenswrapper[4685]: I1013 09:04:26.377488 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.984662281 podStartE2EDuration="7.377475442s" podCreationTimestamp="2025-10-13 09:04:19 +0000 UTC" firstStartedPulling="2025-10-13 09:04:20.797897116 +0000 UTC m=+1185.945772877" lastFinishedPulling="2025-10-13 09:04:25.190710277 +0000 UTC m=+1190.338586038" observedRunningTime="2025-10-13 09:04:26.375736254 +0000 UTC m=+1191.523612005" watchObservedRunningTime="2025-10-13 09:04:26.377475442 +0000 UTC m=+1191.525351203" Oct 13 09:04:26 crc kubenswrapper[4685]: I1013 09:04:26.440431 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.6719800559999998 podStartE2EDuration="7.437595818s" podCreationTimestamp="2025-10-13 09:04:19 +0000 UTC" firstStartedPulling="2025-10-13 09:04:20.430367689 +0000 UTC m=+1185.578243450" lastFinishedPulling="2025-10-13 09:04:25.195983451 +0000 UTC m=+1190.343859212" observedRunningTime="2025-10-13 09:04:26.434491054 +0000 UTC m=+1191.582366815" watchObservedRunningTime="2025-10-13 09:04:26.437595818 +0000 UTC m=+1191.585471579" Oct 13 09:04:26 crc kubenswrapper[4685]: I1013 09:04:26.444491 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.817601352 podStartE2EDuration="7.444481146s" podCreationTimestamp="2025-10-13 09:04:19 +0000 UTC" firstStartedPulling="2025-10-13 09:04:20.561519471 +0000 UTC m=+1185.709395232" lastFinishedPulling="2025-10-13 09:04:25.188399265 +0000 UTC m=+1190.336275026" observedRunningTime="2025-10-13 09:04:26.416625908 +0000 UTC m=+1191.564501669" watchObservedRunningTime="2025-10-13 09:04:26.444481146 +0000 UTC m=+1191.592356907" Oct 13 09:04:26 crc kubenswrapper[4685]: I1013 09:04:26.456859 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.978620116 podStartE2EDuration="7.456844793s" podCreationTimestamp="2025-10-13 09:04:19 +0000 UTC" firstStartedPulling="2025-10-13 09:04:20.710566519 +0000 UTC m=+1185.858442270" lastFinishedPulling="2025-10-13 09:04:25.188791186 +0000 UTC m=+1190.336666947" observedRunningTime="2025-10-13 09:04:26.45598925 +0000 UTC m=+1191.603865031" watchObservedRunningTime="2025-10-13 09:04:26.456844793 +0000 UTC m=+1191.604720554" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.001019 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.045130 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e50b5342-8da0-4176-b977-84860f396352-config-data\") pod \"e50b5342-8da0-4176-b977-84860f396352\" (UID: \"e50b5342-8da0-4176-b977-84860f396352\") " Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.045256 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50b5342-8da0-4176-b977-84860f396352-combined-ca-bundle\") pod \"e50b5342-8da0-4176-b977-84860f396352\" (UID: \"e50b5342-8da0-4176-b977-84860f396352\") " Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.045289 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e50b5342-8da0-4176-b977-84860f396352-logs\") pod \"e50b5342-8da0-4176-b977-84860f396352\" (UID: \"e50b5342-8da0-4176-b977-84860f396352\") " Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.045314 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2n9x\" (UniqueName: \"kubernetes.io/projected/e50b5342-8da0-4176-b977-84860f396352-kube-api-access-z2n9x\") pod \"e50b5342-8da0-4176-b977-84860f396352\" (UID: \"e50b5342-8da0-4176-b977-84860f396352\") " Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.045594 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e50b5342-8da0-4176-b977-84860f396352-logs" (OuterVolumeSpecName: "logs") pod "e50b5342-8da0-4176-b977-84860f396352" (UID: "e50b5342-8da0-4176-b977-84860f396352"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.055212 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e50b5342-8da0-4176-b977-84860f396352-kube-api-access-z2n9x" (OuterVolumeSpecName: "kube-api-access-z2n9x") pod "e50b5342-8da0-4176-b977-84860f396352" (UID: "e50b5342-8da0-4176-b977-84860f396352"). InnerVolumeSpecName "kube-api-access-z2n9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.080824 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e50b5342-8da0-4176-b977-84860f396352-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e50b5342-8da0-4176-b977-84860f396352" (UID: "e50b5342-8da0-4176-b977-84860f396352"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.087198 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e50b5342-8da0-4176-b977-84860f396352-config-data" (OuterVolumeSpecName: "config-data") pod "e50b5342-8da0-4176-b977-84860f396352" (UID: "e50b5342-8da0-4176-b977-84860f396352"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.146312 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2n9x\" (UniqueName: \"kubernetes.io/projected/e50b5342-8da0-4176-b977-84860f396352-kube-api-access-z2n9x\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.146351 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e50b5342-8da0-4176-b977-84860f396352-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.146360 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50b5342-8da0-4176-b977-84860f396352-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.146368 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e50b5342-8da0-4176-b977-84860f396352-logs\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.381004 4685 generic.go:334] "Generic (PLEG): container finished" podID="e50b5342-8da0-4176-b977-84860f396352" containerID="be83a0a70a203abfdf60f9f72920ffb147a6d6e328cff0c6b5ac8f2148750d20" exitCode=0 Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.382087 4685 generic.go:334] "Generic (PLEG): container finished" podID="e50b5342-8da0-4176-b977-84860f396352" containerID="6a833e7e95003fca3f99dfc98ff560809af66bf9f2337ef23aeed7abcdebdd22" exitCode=143 Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.383136 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.387411 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e50b5342-8da0-4176-b977-84860f396352","Type":"ContainerDied","Data":"be83a0a70a203abfdf60f9f72920ffb147a6d6e328cff0c6b5ac8f2148750d20"} Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.387471 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e50b5342-8da0-4176-b977-84860f396352","Type":"ContainerDied","Data":"6a833e7e95003fca3f99dfc98ff560809af66bf9f2337ef23aeed7abcdebdd22"} Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.387491 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e50b5342-8da0-4176-b977-84860f396352","Type":"ContainerDied","Data":"4f0ee6be0519670699bd39a3484ffe05aaa96d4e990ad18c0a04b1fadbea3f44"} Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.387516 4685 scope.go:117] "RemoveContainer" containerID="be83a0a70a203abfdf60f9f72920ffb147a6d6e328cff0c6b5ac8f2148750d20" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.410741 4685 scope.go:117] "RemoveContainer" containerID="6a833e7e95003fca3f99dfc98ff560809af66bf9f2337ef23aeed7abcdebdd22" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.435046 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.446188 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.449095 4685 scope.go:117] "RemoveContainer" containerID="be83a0a70a203abfdf60f9f72920ffb147a6d6e328cff0c6b5ac8f2148750d20" Oct 13 09:04:27 crc kubenswrapper[4685]: E1013 09:04:27.452181 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be83a0a70a203abfdf60f9f72920ffb147a6d6e328cff0c6b5ac8f2148750d20\": container with ID starting with be83a0a70a203abfdf60f9f72920ffb147a6d6e328cff0c6b5ac8f2148750d20 not found: ID does not exist" containerID="be83a0a70a203abfdf60f9f72920ffb147a6d6e328cff0c6b5ac8f2148750d20" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.452237 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be83a0a70a203abfdf60f9f72920ffb147a6d6e328cff0c6b5ac8f2148750d20"} err="failed to get container status \"be83a0a70a203abfdf60f9f72920ffb147a6d6e328cff0c6b5ac8f2148750d20\": rpc error: code = NotFound desc = could not find container \"be83a0a70a203abfdf60f9f72920ffb147a6d6e328cff0c6b5ac8f2148750d20\": container with ID starting with be83a0a70a203abfdf60f9f72920ffb147a6d6e328cff0c6b5ac8f2148750d20 not found: ID does not exist" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.452274 4685 scope.go:117] "RemoveContainer" containerID="6a833e7e95003fca3f99dfc98ff560809af66bf9f2337ef23aeed7abcdebdd22" Oct 13 09:04:27 crc kubenswrapper[4685]: E1013 09:04:27.457292 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a833e7e95003fca3f99dfc98ff560809af66bf9f2337ef23aeed7abcdebdd22\": container with ID starting with 6a833e7e95003fca3f99dfc98ff560809af66bf9f2337ef23aeed7abcdebdd22 not found: ID does not exist" containerID="6a833e7e95003fca3f99dfc98ff560809af66bf9f2337ef23aeed7abcdebdd22" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.457351 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a833e7e95003fca3f99dfc98ff560809af66bf9f2337ef23aeed7abcdebdd22"} err="failed to get container status \"6a833e7e95003fca3f99dfc98ff560809af66bf9f2337ef23aeed7abcdebdd22\": rpc error: code = NotFound desc = could not find container \"6a833e7e95003fca3f99dfc98ff560809af66bf9f2337ef23aeed7abcdebdd22\": container with ID starting with 6a833e7e95003fca3f99dfc98ff560809af66bf9f2337ef23aeed7abcdebdd22 not found: ID does not exist" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.457423 4685 scope.go:117] "RemoveContainer" containerID="be83a0a70a203abfdf60f9f72920ffb147a6d6e328cff0c6b5ac8f2148750d20" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.462428 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be83a0a70a203abfdf60f9f72920ffb147a6d6e328cff0c6b5ac8f2148750d20"} err="failed to get container status \"be83a0a70a203abfdf60f9f72920ffb147a6d6e328cff0c6b5ac8f2148750d20\": rpc error: code = NotFound desc = could not find container \"be83a0a70a203abfdf60f9f72920ffb147a6d6e328cff0c6b5ac8f2148750d20\": container with ID starting with be83a0a70a203abfdf60f9f72920ffb147a6d6e328cff0c6b5ac8f2148750d20 not found: ID does not exist" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.462490 4685 scope.go:117] "RemoveContainer" containerID="6a833e7e95003fca3f99dfc98ff560809af66bf9f2337ef23aeed7abcdebdd22" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.464412 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a833e7e95003fca3f99dfc98ff560809af66bf9f2337ef23aeed7abcdebdd22"} err="failed to get container status \"6a833e7e95003fca3f99dfc98ff560809af66bf9f2337ef23aeed7abcdebdd22\": rpc error: code = NotFound desc = could not find container \"6a833e7e95003fca3f99dfc98ff560809af66bf9f2337ef23aeed7abcdebdd22\": container with ID starting with 6a833e7e95003fca3f99dfc98ff560809af66bf9f2337ef23aeed7abcdebdd22 not found: ID does not exist" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.469848 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 13 09:04:27 crc kubenswrapper[4685]: E1013 09:04:27.471391 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e50b5342-8da0-4176-b977-84860f396352" containerName="nova-metadata-metadata" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.471427 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e50b5342-8da0-4176-b977-84860f396352" containerName="nova-metadata-metadata" Oct 13 09:04:27 crc kubenswrapper[4685]: E1013 09:04:27.471466 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e50b5342-8da0-4176-b977-84860f396352" containerName="nova-metadata-log" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.471481 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e50b5342-8da0-4176-b977-84860f396352" containerName="nova-metadata-log" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.471997 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="e50b5342-8da0-4176-b977-84860f396352" containerName="nova-metadata-metadata" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.472029 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="e50b5342-8da0-4176-b977-84860f396352" containerName="nova-metadata-log" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.473635 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.480880 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.481086 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.488094 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.521103 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e50b5342-8da0-4176-b977-84860f396352" path="/var/lib/kubelet/pods/e50b5342-8da0-4176-b977-84860f396352/volumes" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.654647 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18727d77-8139-4413-ab8a-f6f80fa1d31e-logs\") pod \"nova-metadata-0\" (UID: \"18727d77-8139-4413-ab8a-f6f80fa1d31e\") " pod="openstack/nova-metadata-0" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.655113 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18727d77-8139-4413-ab8a-f6f80fa1d31e-config-data\") pod \"nova-metadata-0\" (UID: \"18727d77-8139-4413-ab8a-f6f80fa1d31e\") " pod="openstack/nova-metadata-0" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.655216 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18727d77-8139-4413-ab8a-f6f80fa1d31e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"18727d77-8139-4413-ab8a-f6f80fa1d31e\") " pod="openstack/nova-metadata-0" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.655255 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9tlx\" (UniqueName: \"kubernetes.io/projected/18727d77-8139-4413-ab8a-f6f80fa1d31e-kube-api-access-c9tlx\") pod \"nova-metadata-0\" (UID: \"18727d77-8139-4413-ab8a-f6f80fa1d31e\") " pod="openstack/nova-metadata-0" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.655282 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/18727d77-8139-4413-ab8a-f6f80fa1d31e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"18727d77-8139-4413-ab8a-f6f80fa1d31e\") " pod="openstack/nova-metadata-0" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.757183 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18727d77-8139-4413-ab8a-f6f80fa1d31e-config-data\") pod \"nova-metadata-0\" (UID: \"18727d77-8139-4413-ab8a-f6f80fa1d31e\") " pod="openstack/nova-metadata-0" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.757259 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18727d77-8139-4413-ab8a-f6f80fa1d31e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"18727d77-8139-4413-ab8a-f6f80fa1d31e\") " pod="openstack/nova-metadata-0" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.757298 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9tlx\" (UniqueName: \"kubernetes.io/projected/18727d77-8139-4413-ab8a-f6f80fa1d31e-kube-api-access-c9tlx\") pod \"nova-metadata-0\" (UID: \"18727d77-8139-4413-ab8a-f6f80fa1d31e\") " pod="openstack/nova-metadata-0" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.757327 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/18727d77-8139-4413-ab8a-f6f80fa1d31e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"18727d77-8139-4413-ab8a-f6f80fa1d31e\") " pod="openstack/nova-metadata-0" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.757393 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18727d77-8139-4413-ab8a-f6f80fa1d31e-logs\") pod \"nova-metadata-0\" (UID: \"18727d77-8139-4413-ab8a-f6f80fa1d31e\") " pod="openstack/nova-metadata-0" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.757755 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18727d77-8139-4413-ab8a-f6f80fa1d31e-logs\") pod \"nova-metadata-0\" (UID: \"18727d77-8139-4413-ab8a-f6f80fa1d31e\") " pod="openstack/nova-metadata-0" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.760995 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18727d77-8139-4413-ab8a-f6f80fa1d31e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"18727d77-8139-4413-ab8a-f6f80fa1d31e\") " pod="openstack/nova-metadata-0" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.761019 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18727d77-8139-4413-ab8a-f6f80fa1d31e-config-data\") pod \"nova-metadata-0\" (UID: \"18727d77-8139-4413-ab8a-f6f80fa1d31e\") " pod="openstack/nova-metadata-0" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.770442 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/18727d77-8139-4413-ab8a-f6f80fa1d31e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"18727d77-8139-4413-ab8a-f6f80fa1d31e\") " pod="openstack/nova-metadata-0" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.785009 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9tlx\" (UniqueName: \"kubernetes.io/projected/18727d77-8139-4413-ab8a-f6f80fa1d31e-kube-api-access-c9tlx\") pod \"nova-metadata-0\" (UID: \"18727d77-8139-4413-ab8a-f6f80fa1d31e\") " pod="openstack/nova-metadata-0" Oct 13 09:04:27 crc kubenswrapper[4685]: I1013 09:04:27.802870 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 09:04:28 crc kubenswrapper[4685]: I1013 09:04:28.289785 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 09:04:28 crc kubenswrapper[4685]: I1013 09:04:28.393326 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"18727d77-8139-4413-ab8a-f6f80fa1d31e","Type":"ContainerStarted","Data":"0ed9231ad36e9608cf2d7db9853b522f3a38a85038d33008fddb10b8287d826a"} Oct 13 09:04:29 crc kubenswrapper[4685]: I1013 09:04:29.412043 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"18727d77-8139-4413-ab8a-f6f80fa1d31e","Type":"ContainerStarted","Data":"f8531792881e7bc29d3cb68fd21a04ed6ce80737477953644267aa7821397659"} Oct 13 09:04:29 crc kubenswrapper[4685]: I1013 09:04:29.412471 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"18727d77-8139-4413-ab8a-f6f80fa1d31e","Type":"ContainerStarted","Data":"1b9d38b64af7f209df3a3eb1e8ea2654eba2ee51f3452f6f21507c4898086a0a"} Oct 13 09:04:29 crc kubenswrapper[4685]: I1013 09:04:29.417145 4685 generic.go:334] "Generic (PLEG): container finished" podID="5c0a4dc6-640e-4aea-9865-aadb943fcc6a" containerID="cbef8ef43372806b6c937531fa6dc0625fe37f464ed56e90a130285aadb2a357" exitCode=0 Oct 13 09:04:29 crc kubenswrapper[4685]: I1013 09:04:29.417206 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5psdz" event={"ID":"5c0a4dc6-640e-4aea-9865-aadb943fcc6a","Type":"ContainerDied","Data":"cbef8ef43372806b6c937531fa6dc0625fe37f464ed56e90a130285aadb2a357"} Oct 13 09:04:29 crc kubenswrapper[4685]: I1013 09:04:29.455743 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.455714269 podStartE2EDuration="2.455714269s" podCreationTimestamp="2025-10-13 09:04:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:04:29.444660737 +0000 UTC m=+1194.592536538" watchObservedRunningTime="2025-10-13 09:04:29.455714269 +0000 UTC m=+1194.603590070" Oct 13 09:04:29 crc kubenswrapper[4685]: I1013 09:04:29.659550 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 13 09:04:29 crc kubenswrapper[4685]: I1013 09:04:29.815814 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 13 09:04:29 crc kubenswrapper[4685]: I1013 09:04:29.815877 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 13 09:04:29 crc kubenswrapper[4685]: I1013 09:04:29.832523 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:29 crc kubenswrapper[4685]: I1013 09:04:29.844736 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 13 09:04:29 crc kubenswrapper[4685]: I1013 09:04:29.855225 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:04:29 crc kubenswrapper[4685]: I1013 09:04:29.878216 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 13 09:04:29 crc kubenswrapper[4685]: I1013 09:04:29.878273 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.001091 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-7c4gj"] Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.001289 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" podUID="6a253ee4-12d7-4e32-9d53-27ecc16ce8cb" containerName="dnsmasq-dns" containerID="cri-o://924dd8afb0dbcb29bcb97540614871285c070d93d517ed44e2aee8133da3bcfd" gracePeriod=10 Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.433198 4685 generic.go:334] "Generic (PLEG): container finished" podID="b8d4a69c-f774-44b5-bb13-5340f07027be" containerID="a110131aaa1c74e0bcd8e70b66e2ec1b6e514ab7428b13aeff24b729d18c411c" exitCode=0 Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.433372 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-wvrpr" event={"ID":"b8d4a69c-f774-44b5-bb13-5340f07027be","Type":"ContainerDied","Data":"a110131aaa1c74e0bcd8e70b66e2ec1b6e514ab7428b13aeff24b729d18c411c"} Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.436445 4685 generic.go:334] "Generic (PLEG): container finished" podID="6a253ee4-12d7-4e32-9d53-27ecc16ce8cb" containerID="924dd8afb0dbcb29bcb97540614871285c070d93d517ed44e2aee8133da3bcfd" exitCode=0 Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.436562 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" event={"ID":"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb","Type":"ContainerDied","Data":"924dd8afb0dbcb29bcb97540614871285c070d93d517ed44e2aee8133da3bcfd"} Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.498846 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.590815 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.750311 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mb6kz\" (UniqueName: \"kubernetes.io/projected/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-kube-api-access-mb6kz\") pod \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.750364 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-dns-swift-storage-0\") pod \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.750417 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-ovsdbserver-sb\") pod \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.750515 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-dns-svc\") pod \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.750622 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-ovsdbserver-nb\") pod \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.750670 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-config\") pod \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\" (UID: \"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb\") " Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.785895 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-kube-api-access-mb6kz" (OuterVolumeSpecName: "kube-api-access-mb6kz") pod "6a253ee4-12d7-4e32-9d53-27ecc16ce8cb" (UID: "6a253ee4-12d7-4e32-9d53-27ecc16ce8cb"). InnerVolumeSpecName "kube-api-access-mb6kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.852837 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mb6kz\" (UniqueName: \"kubernetes.io/projected/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-kube-api-access-mb6kz\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.861713 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6a253ee4-12d7-4e32-9d53-27ecc16ce8cb" (UID: "6a253ee4-12d7-4e32-9d53-27ecc16ce8cb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.874636 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6a253ee4-12d7-4e32-9d53-27ecc16ce8cb" (UID: "6a253ee4-12d7-4e32-9d53-27ecc16ce8cb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.881853 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6a253ee4-12d7-4e32-9d53-27ecc16ce8cb" (UID: "6a253ee4-12d7-4e32-9d53-27ecc16ce8cb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.885345 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5psdz" Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.886341 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6a253ee4-12d7-4e32-9d53-27ecc16ce8cb" (UID: "6a253ee4-12d7-4e32-9d53-27ecc16ce8cb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.926797 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-config" (OuterVolumeSpecName: "config") pod "6a253ee4-12d7-4e32-9d53-27ecc16ce8cb" (UID: "6a253ee4-12d7-4e32-9d53-27ecc16ce8cb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.954871 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.954922 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-config\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.954939 4685 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.954954 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.954964 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.965066 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b00bb156-f7b2-485b-8dce-424707695f9b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.190:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 13 09:04:30 crc kubenswrapper[4685]: I1013 09:04:30.965116 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b00bb156-f7b2-485b-8dce-424707695f9b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.190:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.056525 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-scripts\") pod \"5c0a4dc6-640e-4aea-9865-aadb943fcc6a\" (UID: \"5c0a4dc6-640e-4aea-9865-aadb943fcc6a\") " Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.056670 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-combined-ca-bundle\") pod \"5c0a4dc6-640e-4aea-9865-aadb943fcc6a\" (UID: \"5c0a4dc6-640e-4aea-9865-aadb943fcc6a\") " Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.056722 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvjdx\" (UniqueName: \"kubernetes.io/projected/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-kube-api-access-dvjdx\") pod \"5c0a4dc6-640e-4aea-9865-aadb943fcc6a\" (UID: \"5c0a4dc6-640e-4aea-9865-aadb943fcc6a\") " Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.056752 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-config-data\") pod \"5c0a4dc6-640e-4aea-9865-aadb943fcc6a\" (UID: \"5c0a4dc6-640e-4aea-9865-aadb943fcc6a\") " Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.060098 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-kube-api-access-dvjdx" (OuterVolumeSpecName: "kube-api-access-dvjdx") pod "5c0a4dc6-640e-4aea-9865-aadb943fcc6a" (UID: "5c0a4dc6-640e-4aea-9865-aadb943fcc6a"). InnerVolumeSpecName "kube-api-access-dvjdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.062979 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-scripts" (OuterVolumeSpecName: "scripts") pod "5c0a4dc6-640e-4aea-9865-aadb943fcc6a" (UID: "5c0a4dc6-640e-4aea-9865-aadb943fcc6a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.089726 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-config-data" (OuterVolumeSpecName: "config-data") pod "5c0a4dc6-640e-4aea-9865-aadb943fcc6a" (UID: "5c0a4dc6-640e-4aea-9865-aadb943fcc6a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.092072 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c0a4dc6-640e-4aea-9865-aadb943fcc6a" (UID: "5c0a4dc6-640e-4aea-9865-aadb943fcc6a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.158902 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.158966 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.158980 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvjdx\" (UniqueName: \"kubernetes.io/projected/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-kube-api-access-dvjdx\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.158995 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c0a4dc6-640e-4aea-9865-aadb943fcc6a-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.445098 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5psdz" event={"ID":"5c0a4dc6-640e-4aea-9865-aadb943fcc6a","Type":"ContainerDied","Data":"addc0933f45f0d05018e86dc9981484f4f96fcf6d031da30c2877a7d92e76c69"} Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.445138 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="addc0933f45f0d05018e86dc9981484f4f96fcf6d031da30c2877a7d92e76c69" Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.445157 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5psdz" Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.447986 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" event={"ID":"6a253ee4-12d7-4e32-9d53-27ecc16ce8cb","Type":"ContainerDied","Data":"f755cd6e41eba77fc76e07d259f1af977e371a9faa7beb1a67f52ca4ead82811"} Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.448056 4685 scope.go:117] "RemoveContainer" containerID="924dd8afb0dbcb29bcb97540614871285c070d93d517ed44e2aee8133da3bcfd" Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.448107 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-7c4gj" Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.492227 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-7c4gj"] Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.539400 4685 scope.go:117] "RemoveContainer" containerID="446f23b94fe81a8561329d0ad008a9ac7c533f1b4868061885fb9e65995cb656" Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.551810 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-7c4gj"] Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.625505 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.626067 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b00bb156-f7b2-485b-8dce-424707695f9b" containerName="nova-api-log" containerID="cri-o://f5ddbd46160fd2af94ec224e443d414eb639f90b33562a8997c3bc3dc0103af8" gracePeriod=30 Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.626634 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b00bb156-f7b2-485b-8dce-424707695f9b" containerName="nova-api-api" containerID="cri-o://01f99c9948e44ee05ef7ac3ee3c6de6b135c46be801bf098c2cdd88e13b881b2" gracePeriod=30 Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.671878 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.672140 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="18727d77-8139-4413-ab8a-f6f80fa1d31e" containerName="nova-metadata-log" containerID="cri-o://1b9d38b64af7f209df3a3eb1e8ea2654eba2ee51f3452f6f21507c4898086a0a" gracePeriod=30 Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.672603 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="18727d77-8139-4413-ab8a-f6f80fa1d31e" containerName="nova-metadata-metadata" containerID="cri-o://f8531792881e7bc29d3cb68fd21a04ed6ce80737477953644267aa7821397659" gracePeriod=30 Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.680644 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 09:04:31 crc kubenswrapper[4685]: I1013 09:04:31.942575 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-wvrpr" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.095745 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mshkh\" (UniqueName: \"kubernetes.io/projected/b8d4a69c-f774-44b5-bb13-5340f07027be-kube-api-access-mshkh\") pod \"b8d4a69c-f774-44b5-bb13-5340f07027be\" (UID: \"b8d4a69c-f774-44b5-bb13-5340f07027be\") " Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.095839 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8d4a69c-f774-44b5-bb13-5340f07027be-scripts\") pod \"b8d4a69c-f774-44b5-bb13-5340f07027be\" (UID: \"b8d4a69c-f774-44b5-bb13-5340f07027be\") " Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.095859 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8d4a69c-f774-44b5-bb13-5340f07027be-config-data\") pod \"b8d4a69c-f774-44b5-bb13-5340f07027be\" (UID: \"b8d4a69c-f774-44b5-bb13-5340f07027be\") " Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.096016 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8d4a69c-f774-44b5-bb13-5340f07027be-combined-ca-bundle\") pod \"b8d4a69c-f774-44b5-bb13-5340f07027be\" (UID: \"b8d4a69c-f774-44b5-bb13-5340f07027be\") " Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.108033 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8d4a69c-f774-44b5-bb13-5340f07027be-kube-api-access-mshkh" (OuterVolumeSpecName: "kube-api-access-mshkh") pod "b8d4a69c-f774-44b5-bb13-5340f07027be" (UID: "b8d4a69c-f774-44b5-bb13-5340f07027be"). InnerVolumeSpecName "kube-api-access-mshkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.122561 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8d4a69c-f774-44b5-bb13-5340f07027be-scripts" (OuterVolumeSpecName: "scripts") pod "b8d4a69c-f774-44b5-bb13-5340f07027be" (UID: "b8d4a69c-f774-44b5-bb13-5340f07027be"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.153991 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8d4a69c-f774-44b5-bb13-5340f07027be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b8d4a69c-f774-44b5-bb13-5340f07027be" (UID: "b8d4a69c-f774-44b5-bb13-5340f07027be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.183773 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8d4a69c-f774-44b5-bb13-5340f07027be-config-data" (OuterVolumeSpecName: "config-data") pod "b8d4a69c-f774-44b5-bb13-5340f07027be" (UID: "b8d4a69c-f774-44b5-bb13-5340f07027be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.199079 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8d4a69c-f774-44b5-bb13-5340f07027be-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.199131 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8d4a69c-f774-44b5-bb13-5340f07027be-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.199145 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8d4a69c-f774-44b5-bb13-5340f07027be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.199160 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mshkh\" (UniqueName: \"kubernetes.io/projected/b8d4a69c-f774-44b5-bb13-5340f07027be-kube-api-access-mshkh\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.274473 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.403059 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/18727d77-8139-4413-ab8a-f6f80fa1d31e-nova-metadata-tls-certs\") pod \"18727d77-8139-4413-ab8a-f6f80fa1d31e\" (UID: \"18727d77-8139-4413-ab8a-f6f80fa1d31e\") " Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.403152 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9tlx\" (UniqueName: \"kubernetes.io/projected/18727d77-8139-4413-ab8a-f6f80fa1d31e-kube-api-access-c9tlx\") pod \"18727d77-8139-4413-ab8a-f6f80fa1d31e\" (UID: \"18727d77-8139-4413-ab8a-f6f80fa1d31e\") " Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.403265 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18727d77-8139-4413-ab8a-f6f80fa1d31e-config-data\") pod \"18727d77-8139-4413-ab8a-f6f80fa1d31e\" (UID: \"18727d77-8139-4413-ab8a-f6f80fa1d31e\") " Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.403303 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18727d77-8139-4413-ab8a-f6f80fa1d31e-combined-ca-bundle\") pod \"18727d77-8139-4413-ab8a-f6f80fa1d31e\" (UID: \"18727d77-8139-4413-ab8a-f6f80fa1d31e\") " Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.403325 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18727d77-8139-4413-ab8a-f6f80fa1d31e-logs\") pod \"18727d77-8139-4413-ab8a-f6f80fa1d31e\" (UID: \"18727d77-8139-4413-ab8a-f6f80fa1d31e\") " Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.403996 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18727d77-8139-4413-ab8a-f6f80fa1d31e-logs" (OuterVolumeSpecName: "logs") pod "18727d77-8139-4413-ab8a-f6f80fa1d31e" (UID: "18727d77-8139-4413-ab8a-f6f80fa1d31e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.412038 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18727d77-8139-4413-ab8a-f6f80fa1d31e-kube-api-access-c9tlx" (OuterVolumeSpecName: "kube-api-access-c9tlx") pod "18727d77-8139-4413-ab8a-f6f80fa1d31e" (UID: "18727d77-8139-4413-ab8a-f6f80fa1d31e"). InnerVolumeSpecName "kube-api-access-c9tlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.472479 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18727d77-8139-4413-ab8a-f6f80fa1d31e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18727d77-8139-4413-ab8a-f6f80fa1d31e" (UID: "18727d77-8139-4413-ab8a-f6f80fa1d31e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.476635 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-wvrpr" event={"ID":"b8d4a69c-f774-44b5-bb13-5340f07027be","Type":"ContainerDied","Data":"042aa059d6446d3dfc9f2ba5bf0098b36710242a80840e8e1c241714f46be975"} Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.476673 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="042aa059d6446d3dfc9f2ba5bf0098b36710242a80840e8e1c241714f46be975" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.476724 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-wvrpr" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.507480 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18727d77-8139-4413-ab8a-f6f80fa1d31e-config-data" (OuterVolumeSpecName: "config-data") pod "18727d77-8139-4413-ab8a-f6f80fa1d31e" (UID: "18727d77-8139-4413-ab8a-f6f80fa1d31e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.527138 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18727d77-8139-4413-ab8a-f6f80fa1d31e-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "18727d77-8139-4413-ab8a-f6f80fa1d31e" (UID: "18727d77-8139-4413-ab8a-f6f80fa1d31e"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.527509 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/18727d77-8139-4413-ab8a-f6f80fa1d31e-nova-metadata-tls-certs\") pod \"18727d77-8139-4413-ab8a-f6f80fa1d31e\" (UID: \"18727d77-8139-4413-ab8a-f6f80fa1d31e\") " Oct 13 09:04:32 crc kubenswrapper[4685]: W1013 09:04:32.527598 4685 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/18727d77-8139-4413-ab8a-f6f80fa1d31e/volumes/kubernetes.io~secret/nova-metadata-tls-certs Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.527630 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18727d77-8139-4413-ab8a-f6f80fa1d31e-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "18727d77-8139-4413-ab8a-f6f80fa1d31e" (UID: "18727d77-8139-4413-ab8a-f6f80fa1d31e"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.527512 4685 generic.go:334] "Generic (PLEG): container finished" podID="b00bb156-f7b2-485b-8dce-424707695f9b" containerID="f5ddbd46160fd2af94ec224e443d414eb639f90b33562a8997c3bc3dc0103af8" exitCode=143 Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.527858 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b00bb156-f7b2-485b-8dce-424707695f9b","Type":"ContainerDied","Data":"f5ddbd46160fd2af94ec224e443d414eb639f90b33562a8997c3bc3dc0103af8"} Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.529679 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9tlx\" (UniqueName: \"kubernetes.io/projected/18727d77-8139-4413-ab8a-f6f80fa1d31e-kube-api-access-c9tlx\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.529699 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18727d77-8139-4413-ab8a-f6f80fa1d31e-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.529710 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18727d77-8139-4413-ab8a-f6f80fa1d31e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.529740 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18727d77-8139-4413-ab8a-f6f80fa1d31e-logs\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.529749 4685 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/18727d77-8139-4413-ab8a-f6f80fa1d31e-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.578904 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 13 09:04:32 crc kubenswrapper[4685]: E1013 09:04:32.579327 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8d4a69c-f774-44b5-bb13-5340f07027be" containerName="nova-cell1-conductor-db-sync" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.579341 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8d4a69c-f774-44b5-bb13-5340f07027be" containerName="nova-cell1-conductor-db-sync" Oct 13 09:04:32 crc kubenswrapper[4685]: E1013 09:04:32.579349 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a253ee4-12d7-4e32-9d53-27ecc16ce8cb" containerName="init" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.579354 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a253ee4-12d7-4e32-9d53-27ecc16ce8cb" containerName="init" Oct 13 09:04:32 crc kubenswrapper[4685]: E1013 09:04:32.579370 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a253ee4-12d7-4e32-9d53-27ecc16ce8cb" containerName="dnsmasq-dns" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.579376 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a253ee4-12d7-4e32-9d53-27ecc16ce8cb" containerName="dnsmasq-dns" Oct 13 09:04:32 crc kubenswrapper[4685]: E1013 09:04:32.579399 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18727d77-8139-4413-ab8a-f6f80fa1d31e" containerName="nova-metadata-log" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.579407 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="18727d77-8139-4413-ab8a-f6f80fa1d31e" containerName="nova-metadata-log" Oct 13 09:04:32 crc kubenswrapper[4685]: E1013 09:04:32.579423 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c0a4dc6-640e-4aea-9865-aadb943fcc6a" containerName="nova-manage" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.579429 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c0a4dc6-640e-4aea-9865-aadb943fcc6a" containerName="nova-manage" Oct 13 09:04:32 crc kubenswrapper[4685]: E1013 09:04:32.579442 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18727d77-8139-4413-ab8a-f6f80fa1d31e" containerName="nova-metadata-metadata" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.579447 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="18727d77-8139-4413-ab8a-f6f80fa1d31e" containerName="nova-metadata-metadata" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.579611 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8d4a69c-f774-44b5-bb13-5340f07027be" containerName="nova-cell1-conductor-db-sync" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.579624 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="18727d77-8139-4413-ab8a-f6f80fa1d31e" containerName="nova-metadata-log" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.579641 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c0a4dc6-640e-4aea-9865-aadb943fcc6a" containerName="nova-manage" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.579654 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="18727d77-8139-4413-ab8a-f6f80fa1d31e" containerName="nova-metadata-metadata" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.579669 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a253ee4-12d7-4e32-9d53-27ecc16ce8cb" containerName="dnsmasq-dns" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.580288 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.587335 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.588422 4685 generic.go:334] "Generic (PLEG): container finished" podID="18727d77-8139-4413-ab8a-f6f80fa1d31e" containerID="f8531792881e7bc29d3cb68fd21a04ed6ce80737477953644267aa7821397659" exitCode=0 Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.588435 4685 generic.go:334] "Generic (PLEG): container finished" podID="18727d77-8139-4413-ab8a-f6f80fa1d31e" containerID="1b9d38b64af7f209df3a3eb1e8ea2654eba2ee51f3452f6f21507c4898086a0a" exitCode=143 Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.588667 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="bd3dba3c-5b8e-45a5-a38a-528eea861a6a" containerName="nova-scheduler-scheduler" containerID="cri-o://432d2d8f755a059bf7bb2476e30c65637170774997900033d20f812b8a076a77" gracePeriod=30 Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.588883 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.589035 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"18727d77-8139-4413-ab8a-f6f80fa1d31e","Type":"ContainerDied","Data":"f8531792881e7bc29d3cb68fd21a04ed6ce80737477953644267aa7821397659"} Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.589088 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"18727d77-8139-4413-ab8a-f6f80fa1d31e","Type":"ContainerDied","Data":"1b9d38b64af7f209df3a3eb1e8ea2654eba2ee51f3452f6f21507c4898086a0a"} Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.589101 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"18727d77-8139-4413-ab8a-f6f80fa1d31e","Type":"ContainerDied","Data":"0ed9231ad36e9608cf2d7db9853b522f3a38a85038d33008fddb10b8287d826a"} Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.589118 4685 scope.go:117] "RemoveContainer" containerID="f8531792881e7bc29d3cb68fd21a04ed6ce80737477953644267aa7821397659" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.611515 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.632366 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sktk\" (UniqueName: \"kubernetes.io/projected/b4c0d942-9fdc-4f28-833c-1332cbe23c56-kube-api-access-9sktk\") pod \"nova-cell1-conductor-0\" (UID: \"b4c0d942-9fdc-4f28-833c-1332cbe23c56\") " pod="openstack/nova-cell1-conductor-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.632415 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4c0d942-9fdc-4f28-833c-1332cbe23c56-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b4c0d942-9fdc-4f28-833c-1332cbe23c56\") " pod="openstack/nova-cell1-conductor-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.632479 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4c0d942-9fdc-4f28-833c-1332cbe23c56-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b4c0d942-9fdc-4f28-833c-1332cbe23c56\") " pod="openstack/nova-cell1-conductor-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.719610 4685 scope.go:117] "RemoveContainer" containerID="1b9d38b64af7f209df3a3eb1e8ea2654eba2ee51f3452f6f21507c4898086a0a" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.730202 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.733798 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4c0d942-9fdc-4f28-833c-1332cbe23c56-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b4c0d942-9fdc-4f28-833c-1332cbe23c56\") " pod="openstack/nova-cell1-conductor-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.734009 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4c0d942-9fdc-4f28-833c-1332cbe23c56-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b4c0d942-9fdc-4f28-833c-1332cbe23c56\") " pod="openstack/nova-cell1-conductor-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.734184 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sktk\" (UniqueName: \"kubernetes.io/projected/b4c0d942-9fdc-4f28-833c-1332cbe23c56-kube-api-access-9sktk\") pod \"nova-cell1-conductor-0\" (UID: \"b4c0d942-9fdc-4f28-833c-1332cbe23c56\") " pod="openstack/nova-cell1-conductor-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.741221 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4c0d942-9fdc-4f28-833c-1332cbe23c56-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b4c0d942-9fdc-4f28-833c-1332cbe23c56\") " pod="openstack/nova-cell1-conductor-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.756113 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.756896 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sktk\" (UniqueName: \"kubernetes.io/projected/b4c0d942-9fdc-4f28-833c-1332cbe23c56-kube-api-access-9sktk\") pod \"nova-cell1-conductor-0\" (UID: \"b4c0d942-9fdc-4f28-833c-1332cbe23c56\") " pod="openstack/nova-cell1-conductor-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.757518 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4c0d942-9fdc-4f28-833c-1332cbe23c56-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b4c0d942-9fdc-4f28-833c-1332cbe23c56\") " pod="openstack/nova-cell1-conductor-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.764131 4685 scope.go:117] "RemoveContainer" containerID="f8531792881e7bc29d3cb68fd21a04ed6ce80737477953644267aa7821397659" Oct 13 09:04:32 crc kubenswrapper[4685]: E1013 09:04:32.768227 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8531792881e7bc29d3cb68fd21a04ed6ce80737477953644267aa7821397659\": container with ID starting with f8531792881e7bc29d3cb68fd21a04ed6ce80737477953644267aa7821397659 not found: ID does not exist" containerID="f8531792881e7bc29d3cb68fd21a04ed6ce80737477953644267aa7821397659" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.768266 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8531792881e7bc29d3cb68fd21a04ed6ce80737477953644267aa7821397659"} err="failed to get container status \"f8531792881e7bc29d3cb68fd21a04ed6ce80737477953644267aa7821397659\": rpc error: code = NotFound desc = could not find container \"f8531792881e7bc29d3cb68fd21a04ed6ce80737477953644267aa7821397659\": container with ID starting with f8531792881e7bc29d3cb68fd21a04ed6ce80737477953644267aa7821397659 not found: ID does not exist" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.768288 4685 scope.go:117] "RemoveContainer" containerID="1b9d38b64af7f209df3a3eb1e8ea2654eba2ee51f3452f6f21507c4898086a0a" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.768368 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 13 09:04:32 crc kubenswrapper[4685]: E1013 09:04:32.768710 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b9d38b64af7f209df3a3eb1e8ea2654eba2ee51f3452f6f21507c4898086a0a\": container with ID starting with 1b9d38b64af7f209df3a3eb1e8ea2654eba2ee51f3452f6f21507c4898086a0a not found: ID does not exist" containerID="1b9d38b64af7f209df3a3eb1e8ea2654eba2ee51f3452f6f21507c4898086a0a" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.768842 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b9d38b64af7f209df3a3eb1e8ea2654eba2ee51f3452f6f21507c4898086a0a"} err="failed to get container status \"1b9d38b64af7f209df3a3eb1e8ea2654eba2ee51f3452f6f21507c4898086a0a\": rpc error: code = NotFound desc = could not find container \"1b9d38b64af7f209df3a3eb1e8ea2654eba2ee51f3452f6f21507c4898086a0a\": container with ID starting with 1b9d38b64af7f209df3a3eb1e8ea2654eba2ee51f3452f6f21507c4898086a0a not found: ID does not exist" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.768995 4685 scope.go:117] "RemoveContainer" containerID="f8531792881e7bc29d3cb68fd21a04ed6ce80737477953644267aa7821397659" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.770002 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.770631 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8531792881e7bc29d3cb68fd21a04ed6ce80737477953644267aa7821397659"} err="failed to get container status \"f8531792881e7bc29d3cb68fd21a04ed6ce80737477953644267aa7821397659\": rpc error: code = NotFound desc = could not find container \"f8531792881e7bc29d3cb68fd21a04ed6ce80737477953644267aa7821397659\": container with ID starting with f8531792881e7bc29d3cb68fd21a04ed6ce80737477953644267aa7821397659 not found: ID does not exist" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.770722 4685 scope.go:117] "RemoveContainer" containerID="1b9d38b64af7f209df3a3eb1e8ea2654eba2ee51f3452f6f21507c4898086a0a" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.773499 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.773748 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.773943 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b9d38b64af7f209df3a3eb1e8ea2654eba2ee51f3452f6f21507c4898086a0a"} err="failed to get container status \"1b9d38b64af7f209df3a3eb1e8ea2654eba2ee51f3452f6f21507c4898086a0a\": rpc error: code = NotFound desc = could not find container \"1b9d38b64af7f209df3a3eb1e8ea2654eba2ee51f3452f6f21507c4898086a0a\": container with ID starting with 1b9d38b64af7f209df3a3eb1e8ea2654eba2ee51f3452f6f21507c4898086a0a not found: ID does not exist" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.786313 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.836447 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e61fc8-e032-46a2-822d-3bf80fdf77db-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\") " pod="openstack/nova-metadata-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.836503 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2e61fc8-e032-46a2-822d-3bf80fdf77db-logs\") pod \"nova-metadata-0\" (UID: \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\") " pod="openstack/nova-metadata-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.836714 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2e61fc8-e032-46a2-822d-3bf80fdf77db-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\") " pod="openstack/nova-metadata-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.837054 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7klzs\" (UniqueName: \"kubernetes.io/projected/c2e61fc8-e032-46a2-822d-3bf80fdf77db-kube-api-access-7klzs\") pod \"nova-metadata-0\" (UID: \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\") " pod="openstack/nova-metadata-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.837088 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2e61fc8-e032-46a2-822d-3bf80fdf77db-config-data\") pod \"nova-metadata-0\" (UID: \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\") " pod="openstack/nova-metadata-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.937615 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2e61fc8-e032-46a2-822d-3bf80fdf77db-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\") " pod="openstack/nova-metadata-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.937698 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7klzs\" (UniqueName: \"kubernetes.io/projected/c2e61fc8-e032-46a2-822d-3bf80fdf77db-kube-api-access-7klzs\") pod \"nova-metadata-0\" (UID: \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\") " pod="openstack/nova-metadata-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.937724 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2e61fc8-e032-46a2-822d-3bf80fdf77db-config-data\") pod \"nova-metadata-0\" (UID: \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\") " pod="openstack/nova-metadata-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.937745 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e61fc8-e032-46a2-822d-3bf80fdf77db-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\") " pod="openstack/nova-metadata-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.937762 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2e61fc8-e032-46a2-822d-3bf80fdf77db-logs\") pod \"nova-metadata-0\" (UID: \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\") " pod="openstack/nova-metadata-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.938288 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2e61fc8-e032-46a2-822d-3bf80fdf77db-logs\") pod \"nova-metadata-0\" (UID: \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\") " pod="openstack/nova-metadata-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.942208 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2e61fc8-e032-46a2-822d-3bf80fdf77db-config-data\") pod \"nova-metadata-0\" (UID: \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\") " pod="openstack/nova-metadata-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.942324 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2e61fc8-e032-46a2-822d-3bf80fdf77db-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\") " pod="openstack/nova-metadata-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.942982 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e61fc8-e032-46a2-822d-3bf80fdf77db-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\") " pod="openstack/nova-metadata-0" Oct 13 09:04:32 crc kubenswrapper[4685]: I1013 09:04:32.954828 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7klzs\" (UniqueName: \"kubernetes.io/projected/c2e61fc8-e032-46a2-822d-3bf80fdf77db-kube-api-access-7klzs\") pod \"nova-metadata-0\" (UID: \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\") " pod="openstack/nova-metadata-0" Oct 13 09:04:33 crc kubenswrapper[4685]: I1013 09:04:33.010799 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 13 09:04:33 crc kubenswrapper[4685]: I1013 09:04:33.092956 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 09:04:33 crc kubenswrapper[4685]: I1013 09:04:33.512963 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18727d77-8139-4413-ab8a-f6f80fa1d31e" path="/var/lib/kubelet/pods/18727d77-8139-4413-ab8a-f6f80fa1d31e/volumes" Oct 13 09:04:33 crc kubenswrapper[4685]: I1013 09:04:33.513834 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a253ee4-12d7-4e32-9d53-27ecc16ce8cb" path="/var/lib/kubelet/pods/6a253ee4-12d7-4e32-9d53-27ecc16ce8cb/volumes" Oct 13 09:04:33 crc kubenswrapper[4685]: I1013 09:04:33.550614 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 13 09:04:33 crc kubenswrapper[4685]: I1013 09:04:33.604782 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"b4c0d942-9fdc-4f28-833c-1332cbe23c56","Type":"ContainerStarted","Data":"3f3ff7fe64eac270a61fe51e72be9dd1893c5751513e7ae7e2c8b48a8aee9f7c"} Oct 13 09:04:33 crc kubenswrapper[4685]: I1013 09:04:33.673951 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 09:04:33 crc kubenswrapper[4685]: W1013 09:04:33.689119 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2e61fc8_e032_46a2_822d_3bf80fdf77db.slice/crio-e315ca5134fa7db328cde936ab3c9e03300249c8564811529626feab2f12297f WatchSource:0}: Error finding container e315ca5134fa7db328cde936ab3c9e03300249c8564811529626feab2f12297f: Status 404 returned error can't find the container with id e315ca5134fa7db328cde936ab3c9e03300249c8564811529626feab2f12297f Oct 13 09:04:33 crc kubenswrapper[4685]: I1013 09:04:33.928891 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.070299 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd3dba3c-5b8e-45a5-a38a-528eea861a6a-config-data\") pod \"bd3dba3c-5b8e-45a5-a38a-528eea861a6a\" (UID: \"bd3dba3c-5b8e-45a5-a38a-528eea861a6a\") " Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.070894 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd3dba3c-5b8e-45a5-a38a-528eea861a6a-combined-ca-bundle\") pod \"bd3dba3c-5b8e-45a5-a38a-528eea861a6a\" (UID: \"bd3dba3c-5b8e-45a5-a38a-528eea861a6a\") " Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.071048 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8njl\" (UniqueName: \"kubernetes.io/projected/bd3dba3c-5b8e-45a5-a38a-528eea861a6a-kube-api-access-d8njl\") pod \"bd3dba3c-5b8e-45a5-a38a-528eea861a6a\" (UID: \"bd3dba3c-5b8e-45a5-a38a-528eea861a6a\") " Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.086205 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd3dba3c-5b8e-45a5-a38a-528eea861a6a-kube-api-access-d8njl" (OuterVolumeSpecName: "kube-api-access-d8njl") pod "bd3dba3c-5b8e-45a5-a38a-528eea861a6a" (UID: "bd3dba3c-5b8e-45a5-a38a-528eea861a6a"). InnerVolumeSpecName "kube-api-access-d8njl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.115753 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd3dba3c-5b8e-45a5-a38a-528eea861a6a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd3dba3c-5b8e-45a5-a38a-528eea861a6a" (UID: "bd3dba3c-5b8e-45a5-a38a-528eea861a6a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.117381 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd3dba3c-5b8e-45a5-a38a-528eea861a6a-config-data" (OuterVolumeSpecName: "config-data") pod "bd3dba3c-5b8e-45a5-a38a-528eea861a6a" (UID: "bd3dba3c-5b8e-45a5-a38a-528eea861a6a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.197818 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd3dba3c-5b8e-45a5-a38a-528eea861a6a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.197875 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8njl\" (UniqueName: \"kubernetes.io/projected/bd3dba3c-5b8e-45a5-a38a-528eea861a6a-kube-api-access-d8njl\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.197923 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd3dba3c-5b8e-45a5-a38a-528eea861a6a-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.615149 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"b4c0d942-9fdc-4f28-833c-1332cbe23c56","Type":"ContainerStarted","Data":"8de9d011b0d19253cd9c87569ac1a468d236a98378d89bff4ffed87176e8fea3"} Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.616443 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.617317 4685 generic.go:334] "Generic (PLEG): container finished" podID="bd3dba3c-5b8e-45a5-a38a-528eea861a6a" containerID="432d2d8f755a059bf7bb2476e30c65637170774997900033d20f812b8a076a77" exitCode=0 Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.617412 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.617876 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bd3dba3c-5b8e-45a5-a38a-528eea861a6a","Type":"ContainerDied","Data":"432d2d8f755a059bf7bb2476e30c65637170774997900033d20f812b8a076a77"} Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.617938 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bd3dba3c-5b8e-45a5-a38a-528eea861a6a","Type":"ContainerDied","Data":"62e5efb29fbff7231034cd604144685597d1735bf88ef53cd49f0cd82896dc0f"} Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.618139 4685 scope.go:117] "RemoveContainer" containerID="432d2d8f755a059bf7bb2476e30c65637170774997900033d20f812b8a076a77" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.620295 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2e61fc8-e032-46a2-822d-3bf80fdf77db","Type":"ContainerStarted","Data":"b6c6a2041218515348db8c277deabfd954251d1019139c6e828f0d4b2d0c5ad6"} Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.620328 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2e61fc8-e032-46a2-822d-3bf80fdf77db","Type":"ContainerStarted","Data":"9daae031bca95d2b021c73f80921e5f7417191147f6e574f9d4f4f2fc0c48cc8"} Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.620337 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2e61fc8-e032-46a2-822d-3bf80fdf77db","Type":"ContainerStarted","Data":"e315ca5134fa7db328cde936ab3c9e03300249c8564811529626feab2f12297f"} Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.635299 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.635281582 podStartE2EDuration="2.635281582s" podCreationTimestamp="2025-10-13 09:04:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:04:34.635047986 +0000 UTC m=+1199.782923747" watchObservedRunningTime="2025-10-13 09:04:34.635281582 +0000 UTC m=+1199.783157343" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.649760 4685 scope.go:117] "RemoveContainer" containerID="432d2d8f755a059bf7bb2476e30c65637170774997900033d20f812b8a076a77" Oct 13 09:04:34 crc kubenswrapper[4685]: E1013 09:04:34.654540 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"432d2d8f755a059bf7bb2476e30c65637170774997900033d20f812b8a076a77\": container with ID starting with 432d2d8f755a059bf7bb2476e30c65637170774997900033d20f812b8a076a77 not found: ID does not exist" containerID="432d2d8f755a059bf7bb2476e30c65637170774997900033d20f812b8a076a77" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.654587 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"432d2d8f755a059bf7bb2476e30c65637170774997900033d20f812b8a076a77"} err="failed to get container status \"432d2d8f755a059bf7bb2476e30c65637170774997900033d20f812b8a076a77\": rpc error: code = NotFound desc = could not find container \"432d2d8f755a059bf7bb2476e30c65637170774997900033d20f812b8a076a77\": container with ID starting with 432d2d8f755a059bf7bb2476e30c65637170774997900033d20f812b8a076a77 not found: ID does not exist" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.658038 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.668333 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.680024 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 09:04:34 crc kubenswrapper[4685]: E1013 09:04:34.684295 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd3dba3c-5b8e-45a5-a38a-528eea861a6a" containerName="nova-scheduler-scheduler" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.684428 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd3dba3c-5b8e-45a5-a38a-528eea861a6a" containerName="nova-scheduler-scheduler" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.684789 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd3dba3c-5b8e-45a5-a38a-528eea861a6a" containerName="nova-scheduler-scheduler" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.685534 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.68550811 podStartE2EDuration="2.68550811s" podCreationTimestamp="2025-10-13 09:04:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:04:34.682775726 +0000 UTC m=+1199.830651487" watchObservedRunningTime="2025-10-13 09:04:34.68550811 +0000 UTC m=+1199.833383871" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.688558 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.690495 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.713624 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.808054 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wr9p\" (UniqueName: \"kubernetes.io/projected/306676c4-74d1-4de3-8b41-b16ee72504c7-kube-api-access-7wr9p\") pod \"nova-scheduler-0\" (UID: \"306676c4-74d1-4de3-8b41-b16ee72504c7\") " pod="openstack/nova-scheduler-0" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.808132 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/306676c4-74d1-4de3-8b41-b16ee72504c7-config-data\") pod \"nova-scheduler-0\" (UID: \"306676c4-74d1-4de3-8b41-b16ee72504c7\") " pod="openstack/nova-scheduler-0" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.808208 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/306676c4-74d1-4de3-8b41-b16ee72504c7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"306676c4-74d1-4de3-8b41-b16ee72504c7\") " pod="openstack/nova-scheduler-0" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.912818 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/306676c4-74d1-4de3-8b41-b16ee72504c7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"306676c4-74d1-4de3-8b41-b16ee72504c7\") " pod="openstack/nova-scheduler-0" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.912967 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wr9p\" (UniqueName: \"kubernetes.io/projected/306676c4-74d1-4de3-8b41-b16ee72504c7-kube-api-access-7wr9p\") pod \"nova-scheduler-0\" (UID: \"306676c4-74d1-4de3-8b41-b16ee72504c7\") " pod="openstack/nova-scheduler-0" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.913031 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/306676c4-74d1-4de3-8b41-b16ee72504c7-config-data\") pod \"nova-scheduler-0\" (UID: \"306676c4-74d1-4de3-8b41-b16ee72504c7\") " pod="openstack/nova-scheduler-0" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.918607 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/306676c4-74d1-4de3-8b41-b16ee72504c7-config-data\") pod \"nova-scheduler-0\" (UID: \"306676c4-74d1-4de3-8b41-b16ee72504c7\") " pod="openstack/nova-scheduler-0" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.923800 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/306676c4-74d1-4de3-8b41-b16ee72504c7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"306676c4-74d1-4de3-8b41-b16ee72504c7\") " pod="openstack/nova-scheduler-0" Oct 13 09:04:34 crc kubenswrapper[4685]: I1013 09:04:34.933903 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wr9p\" (UniqueName: \"kubernetes.io/projected/306676c4-74d1-4de3-8b41-b16ee72504c7-kube-api-access-7wr9p\") pod \"nova-scheduler-0\" (UID: \"306676c4-74d1-4de3-8b41-b16ee72504c7\") " pod="openstack/nova-scheduler-0" Oct 13 09:04:35 crc kubenswrapper[4685]: I1013 09:04:35.007875 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 13 09:04:35 crc kubenswrapper[4685]: I1013 09:04:35.539406 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd3dba3c-5b8e-45a5-a38a-528eea861a6a" path="/var/lib/kubelet/pods/bd3dba3c-5b8e-45a5-a38a-528eea861a6a/volumes" Oct 13 09:04:35 crc kubenswrapper[4685]: I1013 09:04:35.554862 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 09:04:35 crc kubenswrapper[4685]: I1013 09:04:35.633170 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"306676c4-74d1-4de3-8b41-b16ee72504c7","Type":"ContainerStarted","Data":"9d77a8548eb8532cb97e2cbc76a2911548749e7caf9631b68549f63da5781f69"} Oct 13 09:04:36 crc kubenswrapper[4685]: I1013 09:04:36.234460 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 13 09:04:36 crc kubenswrapper[4685]: I1013 09:04:36.249830 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="eb9b10e0-0158-4c15-9de6-f8f71ecf531b" containerName="kube-state-metrics" containerID="cri-o://26a6ac5b5aa96d798c66b5a6042ee923909475d47cacf0f29744e11debdcd956" gracePeriod=30 Oct 13 09:04:36 crc kubenswrapper[4685]: E1013 09:04:36.306051 4685 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb9b10e0_0158_4c15_9de6_f8f71ecf531b.slice/crio-conmon-26a6ac5b5aa96d798c66b5a6042ee923909475d47cacf0f29744e11debdcd956.scope\": RecentStats: unable to find data in memory cache]" Oct 13 09:04:36 crc kubenswrapper[4685]: I1013 09:04:36.649206 4685 generic.go:334] "Generic (PLEG): container finished" podID="eb9b10e0-0158-4c15-9de6-f8f71ecf531b" containerID="26a6ac5b5aa96d798c66b5a6042ee923909475d47cacf0f29744e11debdcd956" exitCode=2 Oct 13 09:04:36 crc kubenswrapper[4685]: I1013 09:04:36.649595 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"eb9b10e0-0158-4c15-9de6-f8f71ecf531b","Type":"ContainerDied","Data":"26a6ac5b5aa96d798c66b5a6042ee923909475d47cacf0f29744e11debdcd956"} Oct 13 09:04:36 crc kubenswrapper[4685]: I1013 09:04:36.655902 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"306676c4-74d1-4de3-8b41-b16ee72504c7","Type":"ContainerStarted","Data":"e5a829666a35fab0d360b6c0caf2d2ff2502b0a900fe7e7811827b1d7f1e2224"} Oct 13 09:04:36 crc kubenswrapper[4685]: I1013 09:04:36.678567 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.678543968 podStartE2EDuration="2.678543968s" podCreationTimestamp="2025-10-13 09:04:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:04:36.671816635 +0000 UTC m=+1201.819692396" watchObservedRunningTime="2025-10-13 09:04:36.678543968 +0000 UTC m=+1201.826419749" Oct 13 09:04:36 crc kubenswrapper[4685]: I1013 09:04:36.867003 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.059160 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7q29x\" (UniqueName: \"kubernetes.io/projected/eb9b10e0-0158-4c15-9de6-f8f71ecf531b-kube-api-access-7q29x\") pod \"eb9b10e0-0158-4c15-9de6-f8f71ecf531b\" (UID: \"eb9b10e0-0158-4c15-9de6-f8f71ecf531b\") " Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.066162 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb9b10e0-0158-4c15-9de6-f8f71ecf531b-kube-api-access-7q29x" (OuterVolumeSpecName: "kube-api-access-7q29x") pod "eb9b10e0-0158-4c15-9de6-f8f71ecf531b" (UID: "eb9b10e0-0158-4c15-9de6-f8f71ecf531b"). InnerVolumeSpecName "kube-api-access-7q29x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.161734 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7q29x\" (UniqueName: \"kubernetes.io/projected/eb9b10e0-0158-4c15-9de6-f8f71ecf531b-kube-api-access-7q29x\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.668075 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"eb9b10e0-0158-4c15-9de6-f8f71ecf531b","Type":"ContainerDied","Data":"aba5bbcae489e01ea41a9d026a92a9014669cb501e257e455988cf30e58ce3ee"} Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.668131 4685 scope.go:117] "RemoveContainer" containerID="26a6ac5b5aa96d798c66b5a6042ee923909475d47cacf0f29744e11debdcd956" Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.668185 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.699153 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.716126 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.726475 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 13 09:04:37 crc kubenswrapper[4685]: E1013 09:04:37.727015 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb9b10e0-0158-4c15-9de6-f8f71ecf531b" containerName="kube-state-metrics" Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.727041 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb9b10e0-0158-4c15-9de6-f8f71ecf531b" containerName="kube-state-metrics" Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.727288 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb9b10e0-0158-4c15-9de6-f8f71ecf531b" containerName="kube-state-metrics" Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.727971 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.731962 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.733083 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.770412 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.771555 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a895be7-8da1-4ba2-982a-bde7d7544581-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"6a895be7-8da1-4ba2-982a-bde7d7544581\") " pod="openstack/kube-state-metrics-0" Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.771634 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a895be7-8da1-4ba2-982a-bde7d7544581-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"6a895be7-8da1-4ba2-982a-bde7d7544581\") " pod="openstack/kube-state-metrics-0" Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.771658 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtf5n\" (UniqueName: \"kubernetes.io/projected/6a895be7-8da1-4ba2-982a-bde7d7544581-kube-api-access-jtf5n\") pod \"kube-state-metrics-0\" (UID: \"6a895be7-8da1-4ba2-982a-bde7d7544581\") " pod="openstack/kube-state-metrics-0" Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.771682 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6a895be7-8da1-4ba2-982a-bde7d7544581-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"6a895be7-8da1-4ba2-982a-bde7d7544581\") " pod="openstack/kube-state-metrics-0" Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.872839 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a895be7-8da1-4ba2-982a-bde7d7544581-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"6a895be7-8da1-4ba2-982a-bde7d7544581\") " pod="openstack/kube-state-metrics-0" Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.873345 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a895be7-8da1-4ba2-982a-bde7d7544581-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"6a895be7-8da1-4ba2-982a-bde7d7544581\") " pod="openstack/kube-state-metrics-0" Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.873375 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtf5n\" (UniqueName: \"kubernetes.io/projected/6a895be7-8da1-4ba2-982a-bde7d7544581-kube-api-access-jtf5n\") pod \"kube-state-metrics-0\" (UID: \"6a895be7-8da1-4ba2-982a-bde7d7544581\") " pod="openstack/kube-state-metrics-0" Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.873404 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6a895be7-8da1-4ba2-982a-bde7d7544581-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"6a895be7-8da1-4ba2-982a-bde7d7544581\") " pod="openstack/kube-state-metrics-0" Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.877713 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6a895be7-8da1-4ba2-982a-bde7d7544581-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"6a895be7-8da1-4ba2-982a-bde7d7544581\") " pod="openstack/kube-state-metrics-0" Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.878778 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a895be7-8da1-4ba2-982a-bde7d7544581-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"6a895be7-8da1-4ba2-982a-bde7d7544581\") " pod="openstack/kube-state-metrics-0" Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.883609 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a895be7-8da1-4ba2-982a-bde7d7544581-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"6a895be7-8da1-4ba2-982a-bde7d7544581\") " pod="openstack/kube-state-metrics-0" Oct 13 09:04:37 crc kubenswrapper[4685]: I1013 09:04:37.901520 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtf5n\" (UniqueName: \"kubernetes.io/projected/6a895be7-8da1-4ba2-982a-bde7d7544581-kube-api-access-jtf5n\") pod \"kube-state-metrics-0\" (UID: \"6a895be7-8da1-4ba2-982a-bde7d7544581\") " pod="openstack/kube-state-metrics-0" Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.038573 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.048432 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.094206 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.094597 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.358868 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.359416 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" containerName="proxy-httpd" containerID="cri-o://3d671399559d61833e81b277ebab2cd94594c7316ef9454f073a2cc185dbb6cf" gracePeriod=30 Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.359697 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" containerName="sg-core" containerID="cri-o://5c19ceff41624bd2478d0a35fcb321e096812bbb823fae7adfe826cebd0f2594" gracePeriod=30 Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.360505 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" containerName="ceilometer-notification-agent" containerID="cri-o://6e8bf60e77dda787f80375264cf25c1c9e55e6b418d9a0bfb93bd68d47d9c71d" gracePeriod=30 Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.361601 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" containerName="ceilometer-central-agent" containerID="cri-o://57934e370879cbfe4d820302b9956f0293dbd0ded9ddd5ff96875099bc630935" gracePeriod=30 Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.575895 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.594112 4685 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.606111 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.679997 4685 generic.go:334] "Generic (PLEG): container finished" podID="5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" containerID="3d671399559d61833e81b277ebab2cd94594c7316ef9454f073a2cc185dbb6cf" exitCode=0 Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.680027 4685 generic.go:334] "Generic (PLEG): container finished" podID="5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" containerID="5c19ceff41624bd2478d0a35fcb321e096812bbb823fae7adfe826cebd0f2594" exitCode=2 Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.680053 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103","Type":"ContainerDied","Data":"3d671399559d61833e81b277ebab2cd94594c7316ef9454f073a2cc185dbb6cf"} Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.680093 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103","Type":"ContainerDied","Data":"5c19ceff41624bd2478d0a35fcb321e096812bbb823fae7adfe826cebd0f2594"} Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.681366 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6a895be7-8da1-4ba2-982a-bde7d7544581","Type":"ContainerStarted","Data":"c3c013e3b38f59d45450a115d945ea52a8d51411126c56d30d28d13fab4a1300"} Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.683179 4685 generic.go:334] "Generic (PLEG): container finished" podID="b00bb156-f7b2-485b-8dce-424707695f9b" containerID="01f99c9948e44ee05ef7ac3ee3c6de6b135c46be801bf098c2cdd88e13b881b2" exitCode=0 Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.683233 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.683260 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b00bb156-f7b2-485b-8dce-424707695f9b","Type":"ContainerDied","Data":"01f99c9948e44ee05ef7ac3ee3c6de6b135c46be801bf098c2cdd88e13b881b2"} Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.683293 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b00bb156-f7b2-485b-8dce-424707695f9b","Type":"ContainerDied","Data":"0c4cb5e6cf0dcadeea5dd664332c7af4c9866420f5655e81290e3346861b8741"} Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.683315 4685 scope.go:117] "RemoveContainer" containerID="01f99c9948e44ee05ef7ac3ee3c6de6b135c46be801bf098c2cdd88e13b881b2" Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.705856 4685 scope.go:117] "RemoveContainer" containerID="f5ddbd46160fd2af94ec224e443d414eb639f90b33562a8997c3bc3dc0103af8" Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.735733 4685 scope.go:117] "RemoveContainer" containerID="01f99c9948e44ee05ef7ac3ee3c6de6b135c46be801bf098c2cdd88e13b881b2" Oct 13 09:04:38 crc kubenswrapper[4685]: E1013 09:04:38.740403 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01f99c9948e44ee05ef7ac3ee3c6de6b135c46be801bf098c2cdd88e13b881b2\": container with ID starting with 01f99c9948e44ee05ef7ac3ee3c6de6b135c46be801bf098c2cdd88e13b881b2 not found: ID does not exist" containerID="01f99c9948e44ee05ef7ac3ee3c6de6b135c46be801bf098c2cdd88e13b881b2" Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.740463 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01f99c9948e44ee05ef7ac3ee3c6de6b135c46be801bf098c2cdd88e13b881b2"} err="failed to get container status \"01f99c9948e44ee05ef7ac3ee3c6de6b135c46be801bf098c2cdd88e13b881b2\": rpc error: code = NotFound desc = could not find container \"01f99c9948e44ee05ef7ac3ee3c6de6b135c46be801bf098c2cdd88e13b881b2\": container with ID starting with 01f99c9948e44ee05ef7ac3ee3c6de6b135c46be801bf098c2cdd88e13b881b2 not found: ID does not exist" Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.740491 4685 scope.go:117] "RemoveContainer" containerID="f5ddbd46160fd2af94ec224e443d414eb639f90b33562a8997c3bc3dc0103af8" Oct 13 09:04:38 crc kubenswrapper[4685]: E1013 09:04:38.740847 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5ddbd46160fd2af94ec224e443d414eb639f90b33562a8997c3bc3dc0103af8\": container with ID starting with f5ddbd46160fd2af94ec224e443d414eb639f90b33562a8997c3bc3dc0103af8 not found: ID does not exist" containerID="f5ddbd46160fd2af94ec224e443d414eb639f90b33562a8997c3bc3dc0103af8" Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.740881 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5ddbd46160fd2af94ec224e443d414eb639f90b33562a8997c3bc3dc0103af8"} err="failed to get container status \"f5ddbd46160fd2af94ec224e443d414eb639f90b33562a8997c3bc3dc0103af8\": rpc error: code = NotFound desc = could not find container \"f5ddbd46160fd2af94ec224e443d414eb639f90b33562a8997c3bc3dc0103af8\": container with ID starting with f5ddbd46160fd2af94ec224e443d414eb639f90b33562a8997c3bc3dc0103af8 not found: ID does not exist" Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.794583 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6wtw\" (UniqueName: \"kubernetes.io/projected/b00bb156-f7b2-485b-8dce-424707695f9b-kube-api-access-x6wtw\") pod \"b00bb156-f7b2-485b-8dce-424707695f9b\" (UID: \"b00bb156-f7b2-485b-8dce-424707695f9b\") " Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.794791 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b00bb156-f7b2-485b-8dce-424707695f9b-config-data\") pod \"b00bb156-f7b2-485b-8dce-424707695f9b\" (UID: \"b00bb156-f7b2-485b-8dce-424707695f9b\") " Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.794843 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b00bb156-f7b2-485b-8dce-424707695f9b-logs\") pod \"b00bb156-f7b2-485b-8dce-424707695f9b\" (UID: \"b00bb156-f7b2-485b-8dce-424707695f9b\") " Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.794866 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b00bb156-f7b2-485b-8dce-424707695f9b-combined-ca-bundle\") pod \"b00bb156-f7b2-485b-8dce-424707695f9b\" (UID: \"b00bb156-f7b2-485b-8dce-424707695f9b\") " Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.796409 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b00bb156-f7b2-485b-8dce-424707695f9b-logs" (OuterVolumeSpecName: "logs") pod "b00bb156-f7b2-485b-8dce-424707695f9b" (UID: "b00bb156-f7b2-485b-8dce-424707695f9b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.800679 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b00bb156-f7b2-485b-8dce-424707695f9b-kube-api-access-x6wtw" (OuterVolumeSpecName: "kube-api-access-x6wtw") pod "b00bb156-f7b2-485b-8dce-424707695f9b" (UID: "b00bb156-f7b2-485b-8dce-424707695f9b"). InnerVolumeSpecName "kube-api-access-x6wtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.822300 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b00bb156-f7b2-485b-8dce-424707695f9b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b00bb156-f7b2-485b-8dce-424707695f9b" (UID: "b00bb156-f7b2-485b-8dce-424707695f9b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.825713 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b00bb156-f7b2-485b-8dce-424707695f9b-config-data" (OuterVolumeSpecName: "config-data") pod "b00bb156-f7b2-485b-8dce-424707695f9b" (UID: "b00bb156-f7b2-485b-8dce-424707695f9b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.897150 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b00bb156-f7b2-485b-8dce-424707695f9b-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.897186 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b00bb156-f7b2-485b-8dce-424707695f9b-logs\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.897196 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b00bb156-f7b2-485b-8dce-424707695f9b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:38 crc kubenswrapper[4685]: I1013 09:04:38.897208 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6wtw\" (UniqueName: \"kubernetes.io/projected/b00bb156-f7b2-485b-8dce-424707695f9b-kube-api-access-x6wtw\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.101051 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.110906 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.142340 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 13 09:04:39 crc kubenswrapper[4685]: E1013 09:04:39.144281 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b00bb156-f7b2-485b-8dce-424707695f9b" containerName="nova-api-log" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.144347 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b00bb156-f7b2-485b-8dce-424707695f9b" containerName="nova-api-log" Oct 13 09:04:39 crc kubenswrapper[4685]: E1013 09:04:39.144369 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b00bb156-f7b2-485b-8dce-424707695f9b" containerName="nova-api-api" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.144378 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b00bb156-f7b2-485b-8dce-424707695f9b" containerName="nova-api-api" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.144880 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b00bb156-f7b2-485b-8dce-424707695f9b" containerName="nova-api-api" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.144930 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b00bb156-f7b2-485b-8dce-424707695f9b" containerName="nova-api-log" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.146448 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.152703 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.155425 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.213553 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jbn2\" (UniqueName: \"kubernetes.io/projected/d71cbd40-262f-417f-b535-b35706f88301-kube-api-access-7jbn2\") pod \"nova-api-0\" (UID: \"d71cbd40-262f-417f-b535-b35706f88301\") " pod="openstack/nova-api-0" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.213610 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d71cbd40-262f-417f-b535-b35706f88301-config-data\") pod \"nova-api-0\" (UID: \"d71cbd40-262f-417f-b535-b35706f88301\") " pod="openstack/nova-api-0" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.213647 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d71cbd40-262f-417f-b535-b35706f88301-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d71cbd40-262f-417f-b535-b35706f88301\") " pod="openstack/nova-api-0" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.213689 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d71cbd40-262f-417f-b535-b35706f88301-logs\") pod \"nova-api-0\" (UID: \"d71cbd40-262f-417f-b535-b35706f88301\") " pod="openstack/nova-api-0" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.316221 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d71cbd40-262f-417f-b535-b35706f88301-logs\") pod \"nova-api-0\" (UID: \"d71cbd40-262f-417f-b535-b35706f88301\") " pod="openstack/nova-api-0" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.316655 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jbn2\" (UniqueName: \"kubernetes.io/projected/d71cbd40-262f-417f-b535-b35706f88301-kube-api-access-7jbn2\") pod \"nova-api-0\" (UID: \"d71cbd40-262f-417f-b535-b35706f88301\") " pod="openstack/nova-api-0" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.316689 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d71cbd40-262f-417f-b535-b35706f88301-config-data\") pod \"nova-api-0\" (UID: \"d71cbd40-262f-417f-b535-b35706f88301\") " pod="openstack/nova-api-0" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.316721 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d71cbd40-262f-417f-b535-b35706f88301-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d71cbd40-262f-417f-b535-b35706f88301\") " pod="openstack/nova-api-0" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.316724 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d71cbd40-262f-417f-b535-b35706f88301-logs\") pod \"nova-api-0\" (UID: \"d71cbd40-262f-417f-b535-b35706f88301\") " pod="openstack/nova-api-0" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.322697 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d71cbd40-262f-417f-b535-b35706f88301-config-data\") pod \"nova-api-0\" (UID: \"d71cbd40-262f-417f-b535-b35706f88301\") " pod="openstack/nova-api-0" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.332502 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jbn2\" (UniqueName: \"kubernetes.io/projected/d71cbd40-262f-417f-b535-b35706f88301-kube-api-access-7jbn2\") pod \"nova-api-0\" (UID: \"d71cbd40-262f-417f-b535-b35706f88301\") " pod="openstack/nova-api-0" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.332509 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d71cbd40-262f-417f-b535-b35706f88301-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d71cbd40-262f-417f-b535-b35706f88301\") " pod="openstack/nova-api-0" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.507295 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.513567 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b00bb156-f7b2-485b-8dce-424707695f9b" path="/var/lib/kubelet/pods/b00bb156-f7b2-485b-8dce-424707695f9b/volumes" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.514294 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb9b10e0-0158-4c15-9de6-f8f71ecf531b" path="/var/lib/kubelet/pods/eb9b10e0-0158-4c15-9de6-f8f71ecf531b/volumes" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.704717 4685 generic.go:334] "Generic (PLEG): container finished" podID="5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" containerID="57934e370879cbfe4d820302b9956f0293dbd0ded9ddd5ff96875099bc630935" exitCode=0 Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.705044 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103","Type":"ContainerDied","Data":"57934e370879cbfe4d820302b9956f0293dbd0ded9ddd5ff96875099bc630935"} Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.707045 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6a895be7-8da1-4ba2-982a-bde7d7544581","Type":"ContainerStarted","Data":"74e722d0f7bc05e519c8d061bf7fe608204db04de687ae800c7992b3df0add3b"} Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.707679 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.730854 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.342559607 podStartE2EDuration="2.730832898s" podCreationTimestamp="2025-10-13 09:04:37 +0000 UTC" firstStartedPulling="2025-10-13 09:04:38.593748127 +0000 UTC m=+1203.741623888" lastFinishedPulling="2025-10-13 09:04:38.982021418 +0000 UTC m=+1204.129897179" observedRunningTime="2025-10-13 09:04:39.727634261 +0000 UTC m=+1204.875510042" watchObservedRunningTime="2025-10-13 09:04:39.730832898 +0000 UTC m=+1204.878708679" Oct 13 09:04:39 crc kubenswrapper[4685]: I1013 09:04:39.966816 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 13 09:04:39 crc kubenswrapper[4685]: W1013 09:04:39.971189 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd71cbd40_262f_417f_b535_b35706f88301.slice/crio-c8523b97cb163a116002945485a5f33be7ba98960fe4ea37cbd0bcedebef7083 WatchSource:0}: Error finding container c8523b97cb163a116002945485a5f33be7ba98960fe4ea37cbd0bcedebef7083: Status 404 returned error can't find the container with id c8523b97cb163a116002945485a5f33be7ba98960fe4ea37cbd0bcedebef7083 Oct 13 09:04:40 crc kubenswrapper[4685]: I1013 09:04:40.008471 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 13 09:04:40 crc kubenswrapper[4685]: I1013 09:04:40.723189 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d71cbd40-262f-417f-b535-b35706f88301","Type":"ContainerStarted","Data":"b1a8addace248537cf4bffa053340cfcff4878c7043ddb56a0e3d85dac6112b9"} Oct 13 09:04:40 crc kubenswrapper[4685]: I1013 09:04:40.724393 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d71cbd40-262f-417f-b535-b35706f88301","Type":"ContainerStarted","Data":"d97d01e4fac413c61c4c3c6e22647eb79225a4e2822ed23187fc457f95bf12a8"} Oct 13 09:04:40 crc kubenswrapper[4685]: I1013 09:04:40.724506 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d71cbd40-262f-417f-b535-b35706f88301","Type":"ContainerStarted","Data":"c8523b97cb163a116002945485a5f33be7ba98960fe4ea37cbd0bcedebef7083"} Oct 13 09:04:43 crc kubenswrapper[4685]: I1013 09:04:43.093538 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 13 09:04:43 crc kubenswrapper[4685]: I1013 09:04:43.094081 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 13 09:04:44 crc kubenswrapper[4685]: I1013 09:04:44.107182 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c2e61fc8-e032-46a2-822d-3bf80fdf77db" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 13 09:04:44 crc kubenswrapper[4685]: I1013 09:04:44.107182 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c2e61fc8-e032-46a2-822d-3bf80fdf77db" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.008435 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.039141 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.064615 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=6.064533959 podStartE2EDuration="6.064533959s" podCreationTimestamp="2025-10-13 09:04:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:04:40.745155997 +0000 UTC m=+1205.893031758" watchObservedRunningTime="2025-10-13 09:04:45.064533959 +0000 UTC m=+1210.212409720" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.333548 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.529079 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-scripts\") pod \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.529160 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-log-httpd\") pod \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.529220 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-sg-core-conf-yaml\") pod \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.529320 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-combined-ca-bundle\") pod \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.529339 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-run-httpd\") pod \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.529357 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsdzh\" (UniqueName: \"kubernetes.io/projected/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-kube-api-access-dsdzh\") pod \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.529396 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-config-data\") pod \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\" (UID: \"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103\") " Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.529570 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" (UID: "5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.529738 4685 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.530213 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" (UID: "5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.533860 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-scripts" (OuterVolumeSpecName: "scripts") pod "5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" (UID: "5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.541148 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-kube-api-access-dsdzh" (OuterVolumeSpecName: "kube-api-access-dsdzh") pod "5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" (UID: "5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103"). InnerVolumeSpecName "kube-api-access-dsdzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.569172 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" (UID: "5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.631965 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.631997 4685 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.632009 4685 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.632019 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsdzh\" (UniqueName: \"kubernetes.io/projected/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-kube-api-access-dsdzh\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.634518 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-config-data" (OuterVolumeSpecName: "config-data") pod "5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" (UID: "5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.643615 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" (UID: "5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.734502 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.734557 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.784227 4685 generic.go:334] "Generic (PLEG): container finished" podID="5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" containerID="6e8bf60e77dda787f80375264cf25c1c9e55e6b418d9a0bfb93bd68d47d9c71d" exitCode=0 Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.784304 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.784384 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103","Type":"ContainerDied","Data":"6e8bf60e77dda787f80375264cf25c1c9e55e6b418d9a0bfb93bd68d47d9c71d"} Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.784426 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103","Type":"ContainerDied","Data":"5c93a48c8d47d9cce8cca20a1071120c8b4f76312088cda81f5724c286df84f7"} Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.784448 4685 scope.go:117] "RemoveContainer" containerID="3d671399559d61833e81b277ebab2cd94594c7316ef9454f073a2cc185dbb6cf" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.824600 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.828341 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.836837 4685 scope.go:117] "RemoveContainer" containerID="5c19ceff41624bd2478d0a35fcb321e096812bbb823fae7adfe826cebd0f2594" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.902231 4685 scope.go:117] "RemoveContainer" containerID="6e8bf60e77dda787f80375264cf25c1c9e55e6b418d9a0bfb93bd68d47d9c71d" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.905129 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.915473 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:04:45 crc kubenswrapper[4685]: E1013 09:04:45.916040 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" containerName="ceilometer-notification-agent" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.916062 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" containerName="ceilometer-notification-agent" Oct 13 09:04:45 crc kubenswrapper[4685]: E1013 09:04:45.916107 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" containerName="proxy-httpd" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.916115 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" containerName="proxy-httpd" Oct 13 09:04:45 crc kubenswrapper[4685]: E1013 09:04:45.916143 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" containerName="ceilometer-central-agent" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.916151 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" containerName="ceilometer-central-agent" Oct 13 09:04:45 crc kubenswrapper[4685]: E1013 09:04:45.916184 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" containerName="sg-core" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.916191 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" containerName="sg-core" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.916385 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" containerName="sg-core" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.916399 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" containerName="ceilometer-central-agent" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.916415 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" containerName="proxy-httpd" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.916424 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" containerName="ceilometer-notification-agent" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.918307 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.920843 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.921286 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.921547 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.929681 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.930339 4685 scope.go:117] "RemoveContainer" containerID="57934e370879cbfe4d820302b9956f0293dbd0ded9ddd5ff96875099bc630935" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.960282 4685 scope.go:117] "RemoveContainer" containerID="3d671399559d61833e81b277ebab2cd94594c7316ef9454f073a2cc185dbb6cf" Oct 13 09:04:45 crc kubenswrapper[4685]: E1013 09:04:45.960846 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d671399559d61833e81b277ebab2cd94594c7316ef9454f073a2cc185dbb6cf\": container with ID starting with 3d671399559d61833e81b277ebab2cd94594c7316ef9454f073a2cc185dbb6cf not found: ID does not exist" containerID="3d671399559d61833e81b277ebab2cd94594c7316ef9454f073a2cc185dbb6cf" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.960893 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d671399559d61833e81b277ebab2cd94594c7316ef9454f073a2cc185dbb6cf"} err="failed to get container status \"3d671399559d61833e81b277ebab2cd94594c7316ef9454f073a2cc185dbb6cf\": rpc error: code = NotFound desc = could not find container \"3d671399559d61833e81b277ebab2cd94594c7316ef9454f073a2cc185dbb6cf\": container with ID starting with 3d671399559d61833e81b277ebab2cd94594c7316ef9454f073a2cc185dbb6cf not found: ID does not exist" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.960952 4685 scope.go:117] "RemoveContainer" containerID="5c19ceff41624bd2478d0a35fcb321e096812bbb823fae7adfe826cebd0f2594" Oct 13 09:04:45 crc kubenswrapper[4685]: E1013 09:04:45.961578 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c19ceff41624bd2478d0a35fcb321e096812bbb823fae7adfe826cebd0f2594\": container with ID starting with 5c19ceff41624bd2478d0a35fcb321e096812bbb823fae7adfe826cebd0f2594 not found: ID does not exist" containerID="5c19ceff41624bd2478d0a35fcb321e096812bbb823fae7adfe826cebd0f2594" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.961619 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c19ceff41624bd2478d0a35fcb321e096812bbb823fae7adfe826cebd0f2594"} err="failed to get container status \"5c19ceff41624bd2478d0a35fcb321e096812bbb823fae7adfe826cebd0f2594\": rpc error: code = NotFound desc = could not find container \"5c19ceff41624bd2478d0a35fcb321e096812bbb823fae7adfe826cebd0f2594\": container with ID starting with 5c19ceff41624bd2478d0a35fcb321e096812bbb823fae7adfe826cebd0f2594 not found: ID does not exist" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.961633 4685 scope.go:117] "RemoveContainer" containerID="6e8bf60e77dda787f80375264cf25c1c9e55e6b418d9a0bfb93bd68d47d9c71d" Oct 13 09:04:45 crc kubenswrapper[4685]: E1013 09:04:45.962128 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e8bf60e77dda787f80375264cf25c1c9e55e6b418d9a0bfb93bd68d47d9c71d\": container with ID starting with 6e8bf60e77dda787f80375264cf25c1c9e55e6b418d9a0bfb93bd68d47d9c71d not found: ID does not exist" containerID="6e8bf60e77dda787f80375264cf25c1c9e55e6b418d9a0bfb93bd68d47d9c71d" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.962251 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e8bf60e77dda787f80375264cf25c1c9e55e6b418d9a0bfb93bd68d47d9c71d"} err="failed to get container status \"6e8bf60e77dda787f80375264cf25c1c9e55e6b418d9a0bfb93bd68d47d9c71d\": rpc error: code = NotFound desc = could not find container \"6e8bf60e77dda787f80375264cf25c1c9e55e6b418d9a0bfb93bd68d47d9c71d\": container with ID starting with 6e8bf60e77dda787f80375264cf25c1c9e55e6b418d9a0bfb93bd68d47d9c71d not found: ID does not exist" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.962293 4685 scope.go:117] "RemoveContainer" containerID="57934e370879cbfe4d820302b9956f0293dbd0ded9ddd5ff96875099bc630935" Oct 13 09:04:45 crc kubenswrapper[4685]: E1013 09:04:45.962779 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57934e370879cbfe4d820302b9956f0293dbd0ded9ddd5ff96875099bc630935\": container with ID starting with 57934e370879cbfe4d820302b9956f0293dbd0ded9ddd5ff96875099bc630935 not found: ID does not exist" containerID="57934e370879cbfe4d820302b9956f0293dbd0ded9ddd5ff96875099bc630935" Oct 13 09:04:45 crc kubenswrapper[4685]: I1013 09:04:45.962819 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57934e370879cbfe4d820302b9956f0293dbd0ded9ddd5ff96875099bc630935"} err="failed to get container status \"57934e370879cbfe4d820302b9956f0293dbd0ded9ddd5ff96875099bc630935\": rpc error: code = NotFound desc = could not find container \"57934e370879cbfe4d820302b9956f0293dbd0ded9ddd5ff96875099bc630935\": container with ID starting with 57934e370879cbfe4d820302b9956f0293dbd0ded9ddd5ff96875099bc630935 not found: ID does not exist" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.045608 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbfx9\" (UniqueName: \"kubernetes.io/projected/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-kube-api-access-cbfx9\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.045653 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.045681 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.045796 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.046076 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-config-data\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.046227 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-log-httpd\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.046364 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-scripts\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.046472 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-run-httpd\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.148544 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-run-httpd\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.149860 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbfx9\" (UniqueName: \"kubernetes.io/projected/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-kube-api-access-cbfx9\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.149907 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.149987 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.150078 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.150153 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-config-data\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.151669 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-log-httpd\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.151904 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-scripts\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.149727 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-run-httpd\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.153346 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-log-httpd\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.155167 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.157618 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.157874 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-config-data\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.158393 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.163719 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-scripts\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.187179 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbfx9\" (UniqueName: \"kubernetes.io/projected/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-kube-api-access-cbfx9\") pod \"ceilometer-0\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.249725 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.776956 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:04:46 crc kubenswrapper[4685]: W1013 09:04:46.781890 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3cc6db6_7e61_42b6_9438_d7b764ecd6e6.slice/crio-d46ce6d7f81ef3f24d214732cc69be72457ef89f652ff60e177ee49b85cdadb2 WatchSource:0}: Error finding container d46ce6d7f81ef3f24d214732cc69be72457ef89f652ff60e177ee49b85cdadb2: Status 404 returned error can't find the container with id d46ce6d7f81ef3f24d214732cc69be72457ef89f652ff60e177ee49b85cdadb2 Oct 13 09:04:46 crc kubenswrapper[4685]: I1013 09:04:46.796706 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6","Type":"ContainerStarted","Data":"d46ce6d7f81ef3f24d214732cc69be72457ef89f652ff60e177ee49b85cdadb2"} Oct 13 09:04:47 crc kubenswrapper[4685]: I1013 09:04:47.515580 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103" path="/var/lib/kubelet/pods/5ebeea9f-5d0d-4b7b-b8fc-2baecdfea103/volumes" Oct 13 09:04:47 crc kubenswrapper[4685]: I1013 09:04:47.807322 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6","Type":"ContainerStarted","Data":"838da4fde187c8d9548026cef3235f29c3bc2280b6f9249455a68b3b69073a8d"} Oct 13 09:04:48 crc kubenswrapper[4685]: I1013 09:04:48.068233 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 13 09:04:48 crc kubenswrapper[4685]: I1013 09:04:48.821389 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6","Type":"ContainerStarted","Data":"d795ab83fa6394b50a8bed74288f0c520dd74393d13b09d477e2fe4f8f397625"} Oct 13 09:04:49 crc kubenswrapper[4685]: I1013 09:04:49.515063 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 13 09:04:49 crc kubenswrapper[4685]: I1013 09:04:49.515420 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 13 09:04:49 crc kubenswrapper[4685]: I1013 09:04:49.830762 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6","Type":"ContainerStarted","Data":"be1fe63895ecc8e1eddec47880d44300a0afcab012d0627b3cef1c7279a30d41"} Oct 13 09:04:50 crc kubenswrapper[4685]: I1013 09:04:50.592090 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d71cbd40-262f-417f-b535-b35706f88301" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.200:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 13 09:04:50 crc kubenswrapper[4685]: I1013 09:04:50.592099 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d71cbd40-262f-417f-b535-b35706f88301" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.200:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 13 09:04:51 crc kubenswrapper[4685]: I1013 09:04:51.850490 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6","Type":"ContainerStarted","Data":"6261cf620d190dec08a6dc5ab69579190ded2023218342b94aef28bb896e791a"} Oct 13 09:04:51 crc kubenswrapper[4685]: I1013 09:04:51.851002 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 13 09:04:51 crc kubenswrapper[4685]: I1013 09:04:51.876013 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.29536314 podStartE2EDuration="6.875991626s" podCreationTimestamp="2025-10-13 09:04:45 +0000 UTC" firstStartedPulling="2025-10-13 09:04:46.784528782 +0000 UTC m=+1211.932404553" lastFinishedPulling="2025-10-13 09:04:50.365157278 +0000 UTC m=+1215.513033039" observedRunningTime="2025-10-13 09:04:51.867873315 +0000 UTC m=+1217.015749096" watchObservedRunningTime="2025-10-13 09:04:51.875991626 +0000 UTC m=+1217.023867387" Oct 13 09:04:52 crc kubenswrapper[4685]: I1013 09:04:52.980223 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:04:52 crc kubenswrapper[4685]: I1013 09:04:52.980270 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:04:53 crc kubenswrapper[4685]: I1013 09:04:53.101238 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 13 09:04:53 crc kubenswrapper[4685]: I1013 09:04:53.111550 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 13 09:04:53 crc kubenswrapper[4685]: I1013 09:04:53.114727 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 13 09:04:53 crc kubenswrapper[4685]: I1013 09:04:53.886307 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 13 09:04:56 crc kubenswrapper[4685]: I1013 09:04:56.850982 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:56 crc kubenswrapper[4685]: I1013 09:04:56.905233 4685 generic.go:334] "Generic (PLEG): container finished" podID="0177ea10-0a27-4cfb-99d8-0454b56e1863" containerID="6a2310a77e69619d81f679dd96fae0d10c00fa331514a47bc1fdc54a76fd5665" exitCode=137 Oct 13 09:04:56 crc kubenswrapper[4685]: I1013 09:04:56.905271 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0177ea10-0a27-4cfb-99d8-0454b56e1863","Type":"ContainerDied","Data":"6a2310a77e69619d81f679dd96fae0d10c00fa331514a47bc1fdc54a76fd5665"} Oct 13 09:04:56 crc kubenswrapper[4685]: I1013 09:04:56.905333 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0177ea10-0a27-4cfb-99d8-0454b56e1863","Type":"ContainerDied","Data":"0a03599f7c4c422e3749e32157e261e205839bbdeaf83bd7daf38c74bd58dc10"} Oct 13 09:04:56 crc kubenswrapper[4685]: I1013 09:04:56.905350 4685 scope.go:117] "RemoveContainer" containerID="6a2310a77e69619d81f679dd96fae0d10c00fa331514a47bc1fdc54a76fd5665" Oct 13 09:04:56 crc kubenswrapper[4685]: I1013 09:04:56.905458 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:56 crc kubenswrapper[4685]: I1013 09:04:56.942588 4685 scope.go:117] "RemoveContainer" containerID="6a2310a77e69619d81f679dd96fae0d10c00fa331514a47bc1fdc54a76fd5665" Oct 13 09:04:56 crc kubenswrapper[4685]: E1013 09:04:56.943055 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a2310a77e69619d81f679dd96fae0d10c00fa331514a47bc1fdc54a76fd5665\": container with ID starting with 6a2310a77e69619d81f679dd96fae0d10c00fa331514a47bc1fdc54a76fd5665 not found: ID does not exist" containerID="6a2310a77e69619d81f679dd96fae0d10c00fa331514a47bc1fdc54a76fd5665" Oct 13 09:04:56 crc kubenswrapper[4685]: I1013 09:04:56.943090 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a2310a77e69619d81f679dd96fae0d10c00fa331514a47bc1fdc54a76fd5665"} err="failed to get container status \"6a2310a77e69619d81f679dd96fae0d10c00fa331514a47bc1fdc54a76fd5665\": rpc error: code = NotFound desc = could not find container \"6a2310a77e69619d81f679dd96fae0d10c00fa331514a47bc1fdc54a76fd5665\": container with ID starting with 6a2310a77e69619d81f679dd96fae0d10c00fa331514a47bc1fdc54a76fd5665 not found: ID does not exist" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.004962 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0177ea10-0a27-4cfb-99d8-0454b56e1863-combined-ca-bundle\") pod \"0177ea10-0a27-4cfb-99d8-0454b56e1863\" (UID: \"0177ea10-0a27-4cfb-99d8-0454b56e1863\") " Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.005172 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbvff\" (UniqueName: \"kubernetes.io/projected/0177ea10-0a27-4cfb-99d8-0454b56e1863-kube-api-access-cbvff\") pod \"0177ea10-0a27-4cfb-99d8-0454b56e1863\" (UID: \"0177ea10-0a27-4cfb-99d8-0454b56e1863\") " Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.005251 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0177ea10-0a27-4cfb-99d8-0454b56e1863-config-data\") pod \"0177ea10-0a27-4cfb-99d8-0454b56e1863\" (UID: \"0177ea10-0a27-4cfb-99d8-0454b56e1863\") " Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.014218 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0177ea10-0a27-4cfb-99d8-0454b56e1863-kube-api-access-cbvff" (OuterVolumeSpecName: "kube-api-access-cbvff") pod "0177ea10-0a27-4cfb-99d8-0454b56e1863" (UID: "0177ea10-0a27-4cfb-99d8-0454b56e1863"). InnerVolumeSpecName "kube-api-access-cbvff". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.067859 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0177ea10-0a27-4cfb-99d8-0454b56e1863-config-data" (OuterVolumeSpecName: "config-data") pod "0177ea10-0a27-4cfb-99d8-0454b56e1863" (UID: "0177ea10-0a27-4cfb-99d8-0454b56e1863"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.086150 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0177ea10-0a27-4cfb-99d8-0454b56e1863-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0177ea10-0a27-4cfb-99d8-0454b56e1863" (UID: "0177ea10-0a27-4cfb-99d8-0454b56e1863"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.107499 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbvff\" (UniqueName: \"kubernetes.io/projected/0177ea10-0a27-4cfb-99d8-0454b56e1863-kube-api-access-cbvff\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.107534 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0177ea10-0a27-4cfb-99d8-0454b56e1863-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.107548 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0177ea10-0a27-4cfb-99d8-0454b56e1863-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.241294 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.253009 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.271174 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 13 09:04:57 crc kubenswrapper[4685]: E1013 09:04:57.271517 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0177ea10-0a27-4cfb-99d8-0454b56e1863" containerName="nova-cell1-novncproxy-novncproxy" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.271535 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="0177ea10-0a27-4cfb-99d8-0454b56e1863" containerName="nova-cell1-novncproxy-novncproxy" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.271738 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="0177ea10-0a27-4cfb-99d8-0454b56e1863" containerName="nova-cell1-novncproxy-novncproxy" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.272324 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.275958 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.276083 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.277070 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.286465 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.413435 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/1039fca1-44b0-46d3-8905-daa12aadea65-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1039fca1-44b0-46d3-8905-daa12aadea65\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.413522 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1039fca1-44b0-46d3-8905-daa12aadea65-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1039fca1-44b0-46d3-8905-daa12aadea65\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.413625 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x866v\" (UniqueName: \"kubernetes.io/projected/1039fca1-44b0-46d3-8905-daa12aadea65-kube-api-access-x866v\") pod \"nova-cell1-novncproxy-0\" (UID: \"1039fca1-44b0-46d3-8905-daa12aadea65\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.413664 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/1039fca1-44b0-46d3-8905-daa12aadea65-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1039fca1-44b0-46d3-8905-daa12aadea65\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.413689 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1039fca1-44b0-46d3-8905-daa12aadea65-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1039fca1-44b0-46d3-8905-daa12aadea65\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.514480 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0177ea10-0a27-4cfb-99d8-0454b56e1863" path="/var/lib/kubelet/pods/0177ea10-0a27-4cfb-99d8-0454b56e1863/volumes" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.516046 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/1039fca1-44b0-46d3-8905-daa12aadea65-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1039fca1-44b0-46d3-8905-daa12aadea65\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.516090 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1039fca1-44b0-46d3-8905-daa12aadea65-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1039fca1-44b0-46d3-8905-daa12aadea65\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.516134 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/1039fca1-44b0-46d3-8905-daa12aadea65-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1039fca1-44b0-46d3-8905-daa12aadea65\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.516195 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1039fca1-44b0-46d3-8905-daa12aadea65-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1039fca1-44b0-46d3-8905-daa12aadea65\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.516266 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x866v\" (UniqueName: \"kubernetes.io/projected/1039fca1-44b0-46d3-8905-daa12aadea65-kube-api-access-x866v\") pod \"nova-cell1-novncproxy-0\" (UID: \"1039fca1-44b0-46d3-8905-daa12aadea65\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.522424 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1039fca1-44b0-46d3-8905-daa12aadea65-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1039fca1-44b0-46d3-8905-daa12aadea65\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.524377 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/1039fca1-44b0-46d3-8905-daa12aadea65-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1039fca1-44b0-46d3-8905-daa12aadea65\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.524893 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/1039fca1-44b0-46d3-8905-daa12aadea65-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1039fca1-44b0-46d3-8905-daa12aadea65\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.528434 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1039fca1-44b0-46d3-8905-daa12aadea65-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1039fca1-44b0-46d3-8905-daa12aadea65\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.540797 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x866v\" (UniqueName: \"kubernetes.io/projected/1039fca1-44b0-46d3-8905-daa12aadea65-kube-api-access-x866v\") pod \"nova-cell1-novncproxy-0\" (UID: \"1039fca1-44b0-46d3-8905-daa12aadea65\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:57 crc kubenswrapper[4685]: I1013 09:04:57.594589 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:04:58 crc kubenswrapper[4685]: I1013 09:04:58.066533 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 13 09:04:58 crc kubenswrapper[4685]: W1013 09:04:58.066610 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1039fca1_44b0_46d3_8905_daa12aadea65.slice/crio-10db216af8a439007e36bccead5a4dc6dfcf1e02530cfd7fa47b65b38cbda3f0 WatchSource:0}: Error finding container 10db216af8a439007e36bccead5a4dc6dfcf1e02530cfd7fa47b65b38cbda3f0: Status 404 returned error can't find the container with id 10db216af8a439007e36bccead5a4dc6dfcf1e02530cfd7fa47b65b38cbda3f0 Oct 13 09:04:58 crc kubenswrapper[4685]: I1013 09:04:58.924844 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1039fca1-44b0-46d3-8905-daa12aadea65","Type":"ContainerStarted","Data":"7bb927c5b6d935d97e019750be169dba6497e0536ce5eefd4e50b3ba58a61521"} Oct 13 09:04:58 crc kubenswrapper[4685]: I1013 09:04:58.925267 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1039fca1-44b0-46d3-8905-daa12aadea65","Type":"ContainerStarted","Data":"10db216af8a439007e36bccead5a4dc6dfcf1e02530cfd7fa47b65b38cbda3f0"} Oct 13 09:04:58 crc kubenswrapper[4685]: I1013 09:04:58.948524 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.948503232 podStartE2EDuration="1.948503232s" podCreationTimestamp="2025-10-13 09:04:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:04:58.939719323 +0000 UTC m=+1224.087595094" watchObservedRunningTime="2025-10-13 09:04:58.948503232 +0000 UTC m=+1224.096378993" Oct 13 09:04:59 crc kubenswrapper[4685]: I1013 09:04:59.518111 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 13 09:04:59 crc kubenswrapper[4685]: I1013 09:04:59.519359 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 13 09:04:59 crc kubenswrapper[4685]: I1013 09:04:59.523714 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 13 09:04:59 crc kubenswrapper[4685]: I1013 09:04:59.557974 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 13 09:04:59 crc kubenswrapper[4685]: I1013 09:04:59.934633 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 13 09:04:59 crc kubenswrapper[4685]: I1013 09:04:59.940720 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 13 09:05:00 crc kubenswrapper[4685]: I1013 09:05:00.152679 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-dnzbg"] Oct 13 09:05:00 crc kubenswrapper[4685]: I1013 09:05:00.159169 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:05:00 crc kubenswrapper[4685]: I1013 09:05:00.175971 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-dnzbg"] Oct 13 09:05:00 crc kubenswrapper[4685]: I1013 09:05:00.309075 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-dnzbg\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:05:00 crc kubenswrapper[4685]: I1013 09:05:00.309162 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-dnzbg\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:05:00 crc kubenswrapper[4685]: I1013 09:05:00.309252 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pznk8\" (UniqueName: \"kubernetes.io/projected/dfd5d590-0f0f-4de4-a373-ac9311796783-kube-api-access-pznk8\") pod \"dnsmasq-dns-cd5cbd7b9-dnzbg\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:05:00 crc kubenswrapper[4685]: I1013 09:05:00.309333 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-dnzbg\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:05:00 crc kubenswrapper[4685]: I1013 09:05:00.309393 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-config\") pod \"dnsmasq-dns-cd5cbd7b9-dnzbg\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:05:00 crc kubenswrapper[4685]: I1013 09:05:00.309412 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-dnzbg\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:05:00 crc kubenswrapper[4685]: I1013 09:05:00.410733 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-dnzbg\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:05:00 crc kubenswrapper[4685]: I1013 09:05:00.410795 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-config\") pod \"dnsmasq-dns-cd5cbd7b9-dnzbg\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:05:00 crc kubenswrapper[4685]: I1013 09:05:00.410818 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-dnzbg\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:05:00 crc kubenswrapper[4685]: I1013 09:05:00.410871 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-dnzbg\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:05:00 crc kubenswrapper[4685]: I1013 09:05:00.410891 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-dnzbg\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:05:00 crc kubenswrapper[4685]: I1013 09:05:00.410957 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pznk8\" (UniqueName: \"kubernetes.io/projected/dfd5d590-0f0f-4de4-a373-ac9311796783-kube-api-access-pznk8\") pod \"dnsmasq-dns-cd5cbd7b9-dnzbg\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:05:00 crc kubenswrapper[4685]: I1013 09:05:00.411839 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-dnzbg\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:05:00 crc kubenswrapper[4685]: I1013 09:05:00.411944 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-dnzbg\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:05:00 crc kubenswrapper[4685]: I1013 09:05:00.412111 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-dnzbg\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:05:00 crc kubenswrapper[4685]: I1013 09:05:00.412409 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-config\") pod \"dnsmasq-dns-cd5cbd7b9-dnzbg\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:05:00 crc kubenswrapper[4685]: I1013 09:05:00.412564 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-dnzbg\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:05:00 crc kubenswrapper[4685]: I1013 09:05:00.436968 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pznk8\" (UniqueName: \"kubernetes.io/projected/dfd5d590-0f0f-4de4-a373-ac9311796783-kube-api-access-pznk8\") pod \"dnsmasq-dns-cd5cbd7b9-dnzbg\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:05:00 crc kubenswrapper[4685]: I1013 09:05:00.502092 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:05:01 crc kubenswrapper[4685]: I1013 09:05:01.062972 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-dnzbg"] Oct 13 09:05:01 crc kubenswrapper[4685]: I1013 09:05:01.953969 4685 generic.go:334] "Generic (PLEG): container finished" podID="dfd5d590-0f0f-4de4-a373-ac9311796783" containerID="e05dc373d653e3abdc9fe349a4a33bafda1bb2555a91c854b35b3c4b4325a766" exitCode=0 Oct 13 09:05:01 crc kubenswrapper[4685]: I1013 09:05:01.955449 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" event={"ID":"dfd5d590-0f0f-4de4-a373-ac9311796783","Type":"ContainerDied","Data":"e05dc373d653e3abdc9fe349a4a33bafda1bb2555a91c854b35b3c4b4325a766"} Oct 13 09:05:01 crc kubenswrapper[4685]: I1013 09:05:01.955485 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" event={"ID":"dfd5d590-0f0f-4de4-a373-ac9311796783","Type":"ContainerStarted","Data":"698641b4ca252630d5694da66cb87f513b3864f1f768e5acfa57087956b06b59"} Oct 13 09:05:02 crc kubenswrapper[4685]: I1013 09:05:02.595276 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:05:02 crc kubenswrapper[4685]: I1013 09:05:02.701454 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 13 09:05:02 crc kubenswrapper[4685]: I1013 09:05:02.964463 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" event={"ID":"dfd5d590-0f0f-4de4-a373-ac9311796783","Type":"ContainerStarted","Data":"82f9ced2205498c22984feb3b2c5ac6396a247b6b34b47113c6872c93a04973d"} Oct 13 09:05:02 crc kubenswrapper[4685]: I1013 09:05:02.964530 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d71cbd40-262f-417f-b535-b35706f88301" containerName="nova-api-log" containerID="cri-o://d97d01e4fac413c61c4c3c6e22647eb79225a4e2822ed23187fc457f95bf12a8" gracePeriod=30 Oct 13 09:05:02 crc kubenswrapper[4685]: I1013 09:05:02.964638 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d71cbd40-262f-417f-b535-b35706f88301" containerName="nova-api-api" containerID="cri-o://b1a8addace248537cf4bffa053340cfcff4878c7043ddb56a0e3d85dac6112b9" gracePeriod=30 Oct 13 09:05:03 crc kubenswrapper[4685]: I1013 09:05:03.973846 4685 generic.go:334] "Generic (PLEG): container finished" podID="d71cbd40-262f-417f-b535-b35706f88301" containerID="d97d01e4fac413c61c4c3c6e22647eb79225a4e2822ed23187fc457f95bf12a8" exitCode=143 Oct 13 09:05:03 crc kubenswrapper[4685]: I1013 09:05:03.974440 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d71cbd40-262f-417f-b535-b35706f88301","Type":"ContainerDied","Data":"d97d01e4fac413c61c4c3c6e22647eb79225a4e2822ed23187fc457f95bf12a8"} Oct 13 09:05:03 crc kubenswrapper[4685]: I1013 09:05:03.974613 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:05:05 crc kubenswrapper[4685]: I1013 09:05:05.135074 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" podStartSLOduration=5.135057594 podStartE2EDuration="5.135057594s" podCreationTimestamp="2025-10-13 09:05:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:05:03.000877183 +0000 UTC m=+1228.148752954" watchObservedRunningTime="2025-10-13 09:05:05.135057594 +0000 UTC m=+1230.282933355" Oct 13 09:05:05 crc kubenswrapper[4685]: I1013 09:05:05.142034 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:05:05 crc kubenswrapper[4685]: I1013 09:05:05.142305 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" containerName="ceilometer-central-agent" containerID="cri-o://838da4fde187c8d9548026cef3235f29c3bc2280b6f9249455a68b3b69073a8d" gracePeriod=30 Oct 13 09:05:05 crc kubenswrapper[4685]: I1013 09:05:05.142366 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" containerName="sg-core" containerID="cri-o://be1fe63895ecc8e1eddec47880d44300a0afcab012d0627b3cef1c7279a30d41" gracePeriod=30 Oct 13 09:05:05 crc kubenswrapper[4685]: I1013 09:05:05.142398 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" containerName="ceilometer-notification-agent" containerID="cri-o://d795ab83fa6394b50a8bed74288f0c520dd74393d13b09d477e2fe4f8f397625" gracePeriod=30 Oct 13 09:05:05 crc kubenswrapper[4685]: I1013 09:05:05.142398 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" containerName="proxy-httpd" containerID="cri-o://6261cf620d190dec08a6dc5ab69579190ded2023218342b94aef28bb896e791a" gracePeriod=30 Oct 13 09:05:05 crc kubenswrapper[4685]: I1013 09:05:05.151938 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.201:3000/\": EOF" Oct 13 09:05:05 crc kubenswrapper[4685]: I1013 09:05:05.992887 4685 generic.go:334] "Generic (PLEG): container finished" podID="e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" containerID="6261cf620d190dec08a6dc5ab69579190ded2023218342b94aef28bb896e791a" exitCode=0 Oct 13 09:05:05 crc kubenswrapper[4685]: I1013 09:05:05.992939 4685 generic.go:334] "Generic (PLEG): container finished" podID="e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" containerID="be1fe63895ecc8e1eddec47880d44300a0afcab012d0627b3cef1c7279a30d41" exitCode=2 Oct 13 09:05:05 crc kubenswrapper[4685]: I1013 09:05:05.992950 4685 generic.go:334] "Generic (PLEG): container finished" podID="e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" containerID="838da4fde187c8d9548026cef3235f29c3bc2280b6f9249455a68b3b69073a8d" exitCode=0 Oct 13 09:05:05 crc kubenswrapper[4685]: I1013 09:05:05.992949 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6","Type":"ContainerDied","Data":"6261cf620d190dec08a6dc5ab69579190ded2023218342b94aef28bb896e791a"} Oct 13 09:05:05 crc kubenswrapper[4685]: I1013 09:05:05.992999 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6","Type":"ContainerDied","Data":"be1fe63895ecc8e1eddec47880d44300a0afcab012d0627b3cef1c7279a30d41"} Oct 13 09:05:05 crc kubenswrapper[4685]: I1013 09:05:05.993012 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6","Type":"ContainerDied","Data":"838da4fde187c8d9548026cef3235f29c3bc2280b6f9249455a68b3b69073a8d"} Oct 13 09:05:06 crc kubenswrapper[4685]: I1013 09:05:06.775517 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 09:05:06 crc kubenswrapper[4685]: I1013 09:05:06.936577 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d71cbd40-262f-417f-b535-b35706f88301-combined-ca-bundle\") pod \"d71cbd40-262f-417f-b535-b35706f88301\" (UID: \"d71cbd40-262f-417f-b535-b35706f88301\") " Oct 13 09:05:06 crc kubenswrapper[4685]: I1013 09:05:06.936862 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d71cbd40-262f-417f-b535-b35706f88301-config-data\") pod \"d71cbd40-262f-417f-b535-b35706f88301\" (UID: \"d71cbd40-262f-417f-b535-b35706f88301\") " Oct 13 09:05:06 crc kubenswrapper[4685]: I1013 09:05:06.936998 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d71cbd40-262f-417f-b535-b35706f88301-logs\") pod \"d71cbd40-262f-417f-b535-b35706f88301\" (UID: \"d71cbd40-262f-417f-b535-b35706f88301\") " Oct 13 09:05:06 crc kubenswrapper[4685]: I1013 09:05:06.937093 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jbn2\" (UniqueName: \"kubernetes.io/projected/d71cbd40-262f-417f-b535-b35706f88301-kube-api-access-7jbn2\") pod \"d71cbd40-262f-417f-b535-b35706f88301\" (UID: \"d71cbd40-262f-417f-b535-b35706f88301\") " Oct 13 09:05:06 crc kubenswrapper[4685]: I1013 09:05:06.938941 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d71cbd40-262f-417f-b535-b35706f88301-logs" (OuterVolumeSpecName: "logs") pod "d71cbd40-262f-417f-b535-b35706f88301" (UID: "d71cbd40-262f-417f-b535-b35706f88301"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:05:06 crc kubenswrapper[4685]: I1013 09:05:06.958757 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d71cbd40-262f-417f-b535-b35706f88301-kube-api-access-7jbn2" (OuterVolumeSpecName: "kube-api-access-7jbn2") pod "d71cbd40-262f-417f-b535-b35706f88301" (UID: "d71cbd40-262f-417f-b535-b35706f88301"). InnerVolumeSpecName "kube-api-access-7jbn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:05:06 crc kubenswrapper[4685]: I1013 09:05:06.969656 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d71cbd40-262f-417f-b535-b35706f88301-config-data" (OuterVolumeSpecName: "config-data") pod "d71cbd40-262f-417f-b535-b35706f88301" (UID: "d71cbd40-262f-417f-b535-b35706f88301"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:05:06 crc kubenswrapper[4685]: I1013 09:05:06.995517 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d71cbd40-262f-417f-b535-b35706f88301-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d71cbd40-262f-417f-b535-b35706f88301" (UID: "d71cbd40-262f-417f-b535-b35706f88301"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.022793 4685 generic.go:334] "Generic (PLEG): container finished" podID="d71cbd40-262f-417f-b535-b35706f88301" containerID="b1a8addace248537cf4bffa053340cfcff4878c7043ddb56a0e3d85dac6112b9" exitCode=0 Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.022836 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d71cbd40-262f-417f-b535-b35706f88301","Type":"ContainerDied","Data":"b1a8addace248537cf4bffa053340cfcff4878c7043ddb56a0e3d85dac6112b9"} Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.022860 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d71cbd40-262f-417f-b535-b35706f88301","Type":"ContainerDied","Data":"c8523b97cb163a116002945485a5f33be7ba98960fe4ea37cbd0bcedebef7083"} Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.022875 4685 scope.go:117] "RemoveContainer" containerID="b1a8addace248537cf4bffa053340cfcff4878c7043ddb56a0e3d85dac6112b9" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.023009 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.039893 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d71cbd40-262f-417f-b535-b35706f88301-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.039931 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d71cbd40-262f-417f-b535-b35706f88301-logs\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.039940 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jbn2\" (UniqueName: \"kubernetes.io/projected/d71cbd40-262f-417f-b535-b35706f88301-kube-api-access-7jbn2\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.039950 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d71cbd40-262f-417f-b535-b35706f88301-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.048118 4685 scope.go:117] "RemoveContainer" containerID="d97d01e4fac413c61c4c3c6e22647eb79225a4e2822ed23187fc457f95bf12a8" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.103697 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.111104 4685 scope.go:117] "RemoveContainer" containerID="b1a8addace248537cf4bffa053340cfcff4878c7043ddb56a0e3d85dac6112b9" Oct 13 09:05:07 crc kubenswrapper[4685]: E1013 09:05:07.112693 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1a8addace248537cf4bffa053340cfcff4878c7043ddb56a0e3d85dac6112b9\": container with ID starting with b1a8addace248537cf4bffa053340cfcff4878c7043ddb56a0e3d85dac6112b9 not found: ID does not exist" containerID="b1a8addace248537cf4bffa053340cfcff4878c7043ddb56a0e3d85dac6112b9" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.112747 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1a8addace248537cf4bffa053340cfcff4878c7043ddb56a0e3d85dac6112b9"} err="failed to get container status \"b1a8addace248537cf4bffa053340cfcff4878c7043ddb56a0e3d85dac6112b9\": rpc error: code = NotFound desc = could not find container \"b1a8addace248537cf4bffa053340cfcff4878c7043ddb56a0e3d85dac6112b9\": container with ID starting with b1a8addace248537cf4bffa053340cfcff4878c7043ddb56a0e3d85dac6112b9 not found: ID does not exist" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.112777 4685 scope.go:117] "RemoveContainer" containerID="d97d01e4fac413c61c4c3c6e22647eb79225a4e2822ed23187fc457f95bf12a8" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.114847 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 13 09:05:07 crc kubenswrapper[4685]: E1013 09:05:07.115545 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d97d01e4fac413c61c4c3c6e22647eb79225a4e2822ed23187fc457f95bf12a8\": container with ID starting with d97d01e4fac413c61c4c3c6e22647eb79225a4e2822ed23187fc457f95bf12a8 not found: ID does not exist" containerID="d97d01e4fac413c61c4c3c6e22647eb79225a4e2822ed23187fc457f95bf12a8" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.115584 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d97d01e4fac413c61c4c3c6e22647eb79225a4e2822ed23187fc457f95bf12a8"} err="failed to get container status \"d97d01e4fac413c61c4c3c6e22647eb79225a4e2822ed23187fc457f95bf12a8\": rpc error: code = NotFound desc = could not find container \"d97d01e4fac413c61c4c3c6e22647eb79225a4e2822ed23187fc457f95bf12a8\": container with ID starting with d97d01e4fac413c61c4c3c6e22647eb79225a4e2822ed23187fc457f95bf12a8 not found: ID does not exist" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.129146 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 13 09:05:07 crc kubenswrapper[4685]: E1013 09:05:07.129586 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d71cbd40-262f-417f-b535-b35706f88301" containerName="nova-api-log" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.129602 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="d71cbd40-262f-417f-b535-b35706f88301" containerName="nova-api-log" Oct 13 09:05:07 crc kubenswrapper[4685]: E1013 09:05:07.129633 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d71cbd40-262f-417f-b535-b35706f88301" containerName="nova-api-api" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.129640 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="d71cbd40-262f-417f-b535-b35706f88301" containerName="nova-api-api" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.129851 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="d71cbd40-262f-417f-b535-b35706f88301" containerName="nova-api-api" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.129876 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="d71cbd40-262f-417f-b535-b35706f88301" containerName="nova-api-log" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.130883 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.137341 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.138050 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.137958 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.138799 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.244668 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-config-data\") pod \"nova-api-0\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " pod="openstack/nova-api-0" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.245031 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-public-tls-certs\") pod \"nova-api-0\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " pod="openstack/nova-api-0" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.245142 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5931019-0ca3-4483-a465-a7a9fc674179-logs\") pod \"nova-api-0\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " pod="openstack/nova-api-0" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.245243 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " pod="openstack/nova-api-0" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.245410 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25j9c\" (UniqueName: \"kubernetes.io/projected/d5931019-0ca3-4483-a465-a7a9fc674179-kube-api-access-25j9c\") pod \"nova-api-0\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " pod="openstack/nova-api-0" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.245515 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " pod="openstack/nova-api-0" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.347315 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-config-data\") pod \"nova-api-0\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " pod="openstack/nova-api-0" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.347371 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-public-tls-certs\") pod \"nova-api-0\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " pod="openstack/nova-api-0" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.347387 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5931019-0ca3-4483-a465-a7a9fc674179-logs\") pod \"nova-api-0\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " pod="openstack/nova-api-0" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.347403 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " pod="openstack/nova-api-0" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.347421 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25j9c\" (UniqueName: \"kubernetes.io/projected/d5931019-0ca3-4483-a465-a7a9fc674179-kube-api-access-25j9c\") pod \"nova-api-0\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " pod="openstack/nova-api-0" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.347445 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " pod="openstack/nova-api-0" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.348522 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5931019-0ca3-4483-a465-a7a9fc674179-logs\") pod \"nova-api-0\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " pod="openstack/nova-api-0" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.352855 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-config-data\") pod \"nova-api-0\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " pod="openstack/nova-api-0" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.353186 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " pod="openstack/nova-api-0" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.355126 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-public-tls-certs\") pod \"nova-api-0\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " pod="openstack/nova-api-0" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.369408 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25j9c\" (UniqueName: \"kubernetes.io/projected/d5931019-0ca3-4483-a465-a7a9fc674179-kube-api-access-25j9c\") pod \"nova-api-0\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " pod="openstack/nova-api-0" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.369618 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " pod="openstack/nova-api-0" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.486553 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.513133 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d71cbd40-262f-417f-b535-b35706f88301" path="/var/lib/kubelet/pods/d71cbd40-262f-417f-b535-b35706f88301/volumes" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.595570 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.634353 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:05:07 crc kubenswrapper[4685]: W1013 09:05:07.988253 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5931019_0ca3_4483_a465_a7a9fc674179.slice/crio-dfd8b98c6008d550f82ab5ce3828fc6da536fa3e6a5a6c89cf856e6e36441aae WatchSource:0}: Error finding container dfd8b98c6008d550f82ab5ce3828fc6da536fa3e6a5a6c89cf856e6e36441aae: Status 404 returned error can't find the container with id dfd8b98c6008d550f82ab5ce3828fc6da536fa3e6a5a6c89cf856e6e36441aae Oct 13 09:05:07 crc kubenswrapper[4685]: I1013 09:05:07.989816 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 13 09:05:08 crc kubenswrapper[4685]: I1013 09:05:08.038473 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d5931019-0ca3-4483-a465-a7a9fc674179","Type":"ContainerStarted","Data":"dfd8b98c6008d550f82ab5ce3828fc6da536fa3e6a5a6c89cf856e6e36441aae"} Oct 13 09:05:08 crc kubenswrapper[4685]: I1013 09:05:08.062425 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 13 09:05:08 crc kubenswrapper[4685]: I1013 09:05:08.345225 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-tkkqw"] Oct 13 09:05:08 crc kubenswrapper[4685]: I1013 09:05:08.347905 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-tkkqw" Oct 13 09:05:08 crc kubenswrapper[4685]: I1013 09:05:08.355491 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 13 09:05:08 crc kubenswrapper[4685]: I1013 09:05:08.355575 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 13 09:05:08 crc kubenswrapper[4685]: I1013 09:05:08.360335 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-tkkqw"] Oct 13 09:05:08 crc kubenswrapper[4685]: I1013 09:05:08.464741 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fae68cb4-7110-40c0-b48b-b38036250c61-config-data\") pod \"nova-cell1-cell-mapping-tkkqw\" (UID: \"fae68cb4-7110-40c0-b48b-b38036250c61\") " pod="openstack/nova-cell1-cell-mapping-tkkqw" Oct 13 09:05:08 crc kubenswrapper[4685]: I1013 09:05:08.465139 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvflh\" (UniqueName: \"kubernetes.io/projected/fae68cb4-7110-40c0-b48b-b38036250c61-kube-api-access-jvflh\") pod \"nova-cell1-cell-mapping-tkkqw\" (UID: \"fae68cb4-7110-40c0-b48b-b38036250c61\") " pod="openstack/nova-cell1-cell-mapping-tkkqw" Oct 13 09:05:08 crc kubenswrapper[4685]: I1013 09:05:08.465168 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fae68cb4-7110-40c0-b48b-b38036250c61-scripts\") pod \"nova-cell1-cell-mapping-tkkqw\" (UID: \"fae68cb4-7110-40c0-b48b-b38036250c61\") " pod="openstack/nova-cell1-cell-mapping-tkkqw" Oct 13 09:05:08 crc kubenswrapper[4685]: I1013 09:05:08.465236 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fae68cb4-7110-40c0-b48b-b38036250c61-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-tkkqw\" (UID: \"fae68cb4-7110-40c0-b48b-b38036250c61\") " pod="openstack/nova-cell1-cell-mapping-tkkqw" Oct 13 09:05:08 crc kubenswrapper[4685]: I1013 09:05:08.567725 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fae68cb4-7110-40c0-b48b-b38036250c61-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-tkkqw\" (UID: \"fae68cb4-7110-40c0-b48b-b38036250c61\") " pod="openstack/nova-cell1-cell-mapping-tkkqw" Oct 13 09:05:08 crc kubenswrapper[4685]: I1013 09:05:08.567994 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fae68cb4-7110-40c0-b48b-b38036250c61-config-data\") pod \"nova-cell1-cell-mapping-tkkqw\" (UID: \"fae68cb4-7110-40c0-b48b-b38036250c61\") " pod="openstack/nova-cell1-cell-mapping-tkkqw" Oct 13 09:05:08 crc kubenswrapper[4685]: I1013 09:05:08.568189 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvflh\" (UniqueName: \"kubernetes.io/projected/fae68cb4-7110-40c0-b48b-b38036250c61-kube-api-access-jvflh\") pod \"nova-cell1-cell-mapping-tkkqw\" (UID: \"fae68cb4-7110-40c0-b48b-b38036250c61\") " pod="openstack/nova-cell1-cell-mapping-tkkqw" Oct 13 09:05:08 crc kubenswrapper[4685]: I1013 09:05:08.568281 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fae68cb4-7110-40c0-b48b-b38036250c61-scripts\") pod \"nova-cell1-cell-mapping-tkkqw\" (UID: \"fae68cb4-7110-40c0-b48b-b38036250c61\") " pod="openstack/nova-cell1-cell-mapping-tkkqw" Oct 13 09:05:08 crc kubenswrapper[4685]: I1013 09:05:08.574736 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fae68cb4-7110-40c0-b48b-b38036250c61-scripts\") pod \"nova-cell1-cell-mapping-tkkqw\" (UID: \"fae68cb4-7110-40c0-b48b-b38036250c61\") " pod="openstack/nova-cell1-cell-mapping-tkkqw" Oct 13 09:05:08 crc kubenswrapper[4685]: I1013 09:05:08.574759 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fae68cb4-7110-40c0-b48b-b38036250c61-config-data\") pod \"nova-cell1-cell-mapping-tkkqw\" (UID: \"fae68cb4-7110-40c0-b48b-b38036250c61\") " pod="openstack/nova-cell1-cell-mapping-tkkqw" Oct 13 09:05:08 crc kubenswrapper[4685]: I1013 09:05:08.575248 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fae68cb4-7110-40c0-b48b-b38036250c61-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-tkkqw\" (UID: \"fae68cb4-7110-40c0-b48b-b38036250c61\") " pod="openstack/nova-cell1-cell-mapping-tkkqw" Oct 13 09:05:08 crc kubenswrapper[4685]: I1013 09:05:08.591378 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvflh\" (UniqueName: \"kubernetes.io/projected/fae68cb4-7110-40c0-b48b-b38036250c61-kube-api-access-jvflh\") pod \"nova-cell1-cell-mapping-tkkqw\" (UID: \"fae68cb4-7110-40c0-b48b-b38036250c61\") " pod="openstack/nova-cell1-cell-mapping-tkkqw" Oct 13 09:05:08 crc kubenswrapper[4685]: I1013 09:05:08.686250 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-tkkqw" Oct 13 09:05:09 crc kubenswrapper[4685]: I1013 09:05:09.063084 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d5931019-0ca3-4483-a465-a7a9fc674179","Type":"ContainerStarted","Data":"79ea6cf83c728649d57afbd4d598c450785086141be3dc578e47d9f5ea8a6414"} Oct 13 09:05:09 crc kubenswrapper[4685]: I1013 09:05:09.063501 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d5931019-0ca3-4483-a465-a7a9fc674179","Type":"ContainerStarted","Data":"d8134cc82308c9452bed5425b7c4dab52375247bb537ed439f0b98fc2de3bf88"} Oct 13 09:05:09 crc kubenswrapper[4685]: I1013 09:05:09.089728 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.089710343 podStartE2EDuration="2.089710343s" podCreationTimestamp="2025-10-13 09:05:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:05:09.084765649 +0000 UTC m=+1234.232641430" watchObservedRunningTime="2025-10-13 09:05:09.089710343 +0000 UTC m=+1234.237586104" Oct 13 09:05:09 crc kubenswrapper[4685]: I1013 09:05:09.219030 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-tkkqw"] Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.077086 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-tkkqw" event={"ID":"fae68cb4-7110-40c0-b48b-b38036250c61","Type":"ContainerStarted","Data":"4387f4f578b34308748b28e4f9426909a73eca7b00aaa8820ff473823a37df64"} Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.077594 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-tkkqw" event={"ID":"fae68cb4-7110-40c0-b48b-b38036250c61","Type":"ContainerStarted","Data":"868fd71d1609daa2498ee9dcd81aaff3726eb2f2171ce0bb7b20744b20b865e2"} Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.087797 4685 generic.go:334] "Generic (PLEG): container finished" podID="e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" containerID="d795ab83fa6394b50a8bed74288f0c520dd74393d13b09d477e2fe4f8f397625" exitCode=0 Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.087870 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6","Type":"ContainerDied","Data":"d795ab83fa6394b50a8bed74288f0c520dd74393d13b09d477e2fe4f8f397625"} Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.098249 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-tkkqw" podStartSLOduration=2.098234239 podStartE2EDuration="2.098234239s" podCreationTimestamp="2025-10-13 09:05:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:05:10.096137572 +0000 UTC m=+1235.244013333" watchObservedRunningTime="2025-10-13 09:05:10.098234239 +0000 UTC m=+1235.246110000" Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.462649 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.504060 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.535976 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-sg-core-conf-yaml\") pod \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.536762 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-run-httpd\") pod \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.536786 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-scripts\") pod \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.537082 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbfx9\" (UniqueName: \"kubernetes.io/projected/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-kube-api-access-cbfx9\") pod \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.537133 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-log-httpd\") pod \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.537191 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" (UID: "e3cc6db6-7e61-42b6-9438-d7b764ecd6e6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.537291 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-ceilometer-tls-certs\") pod \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.537427 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-config-data\") pod \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.537461 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-combined-ca-bundle\") pod \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\" (UID: \"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6\") " Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.538560 4685 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.539260 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" (UID: "e3cc6db6-7e61-42b6-9438-d7b764ecd6e6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.569760 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-kube-api-access-cbfx9" (OuterVolumeSpecName: "kube-api-access-cbfx9") pod "e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" (UID: "e3cc6db6-7e61-42b6-9438-d7b764ecd6e6"). InnerVolumeSpecName "kube-api-access-cbfx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.581557 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-scripts" (OuterVolumeSpecName: "scripts") pod "e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" (UID: "e3cc6db6-7e61-42b6-9438-d7b764ecd6e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.601825 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" (UID: "e3cc6db6-7e61-42b6-9438-d7b764ecd6e6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.631384 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" (UID: "e3cc6db6-7e61-42b6-9438-d7b764ecd6e6"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.647996 4685 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.648023 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.648032 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbfx9\" (UniqueName: \"kubernetes.io/projected/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-kube-api-access-cbfx9\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.648040 4685 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.648049 4685 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.660437 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-nv9ck"] Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.660694 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" podUID="7f5ec6ac-31a4-4d32-a600-2c274d9535df" containerName="dnsmasq-dns" containerID="cri-o://7ce9dadc75ad7f16a641fb2b4705e647595b8b2a879975af4bb44706eca7143e" gracePeriod=10 Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.788337 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" (UID: "e3cc6db6-7e61-42b6-9438-d7b764ecd6e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.855211 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.949474 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-config-data" (OuterVolumeSpecName: "config-data") pod "e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" (UID: "e3cc6db6-7e61-42b6-9438-d7b764ecd6e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:05:10 crc kubenswrapper[4685]: I1013 09:05:10.957414 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.096843 4685 generic.go:334] "Generic (PLEG): container finished" podID="7f5ec6ac-31a4-4d32-a600-2c274d9535df" containerID="7ce9dadc75ad7f16a641fb2b4705e647595b8b2a879975af4bb44706eca7143e" exitCode=0 Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.096899 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" event={"ID":"7f5ec6ac-31a4-4d32-a600-2c274d9535df","Type":"ContainerDied","Data":"7ce9dadc75ad7f16a641fb2b4705e647595b8b2a879975af4bb44706eca7143e"} Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.098882 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3cc6db6-7e61-42b6-9438-d7b764ecd6e6","Type":"ContainerDied","Data":"d46ce6d7f81ef3f24d214732cc69be72457ef89f652ff60e177ee49b85cdadb2"} Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.099070 4685 scope.go:117] "RemoveContainer" containerID="6261cf620d190dec08a6dc5ab69579190ded2023218342b94aef28bb896e791a" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.098942 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.121714 4685 scope.go:117] "RemoveContainer" containerID="be1fe63895ecc8e1eddec47880d44300a0afcab012d0627b3cef1c7279a30d41" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.141298 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.142262 4685 scope.go:117] "RemoveContainer" containerID="d795ab83fa6394b50a8bed74288f0c520dd74393d13b09d477e2fe4f8f397625" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.158413 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.181429 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:05:11 crc kubenswrapper[4685]: E1013 09:05:11.181890 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" containerName="sg-core" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.181903 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" containerName="sg-core" Oct 13 09:05:11 crc kubenswrapper[4685]: E1013 09:05:11.181936 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" containerName="ceilometer-notification-agent" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.181943 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" containerName="ceilometer-notification-agent" Oct 13 09:05:11 crc kubenswrapper[4685]: E1013 09:05:11.181956 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" containerName="proxy-httpd" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.181962 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" containerName="proxy-httpd" Oct 13 09:05:11 crc kubenswrapper[4685]: E1013 09:05:11.181969 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" containerName="ceilometer-central-agent" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.181976 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" containerName="ceilometer-central-agent" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.182188 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" containerName="sg-core" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.182207 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" containerName="ceilometer-notification-agent" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.182221 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" containerName="ceilometer-central-agent" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.182228 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" containerName="proxy-httpd" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.183817 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.186435 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.187334 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.187934 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.196299 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.230256 4685 scope.go:117] "RemoveContainer" containerID="838da4fde187c8d9548026cef3235f29c3bc2280b6f9249455a68b3b69073a8d" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.263729 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-scripts\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.264047 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.264178 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.264266 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prbgt\" (UniqueName: \"kubernetes.io/projected/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-kube-api-access-prbgt\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.264336 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-config-data\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.264410 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.264482 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-log-httpd\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.264573 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-run-httpd\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.366313 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.366407 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prbgt\" (UniqueName: \"kubernetes.io/projected/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-kube-api-access-prbgt\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.366451 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-config-data\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.366481 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.366508 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-log-httpd\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.366556 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-run-httpd\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.366574 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-scripts\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.366667 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.368310 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-run-httpd\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.368339 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-log-httpd\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.371016 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-scripts\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.382704 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.382834 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-config-data\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.383373 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.389807 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prbgt\" (UniqueName: \"kubernetes.io/projected/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-kube-api-access-prbgt\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.391722 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5aa884f8-807f-407b-9ea0-ea0812fcd1a2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5aa884f8-807f-407b-9ea0-ea0812fcd1a2\") " pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.506474 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.517959 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3cc6db6-7e61-42b6-9438-d7b764ecd6e6" path="/var/lib/kubelet/pods/e3cc6db6-7e61-42b6-9438-d7b764ecd6e6/volumes" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.611456 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.673159 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-dns-swift-storage-0\") pod \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.673228 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-config\") pod \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.673286 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-ovsdbserver-nb\") pod \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.673328 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-ovsdbserver-sb\") pod \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.673375 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-dns-svc\") pod \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.673463 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwvs8\" (UniqueName: \"kubernetes.io/projected/7f5ec6ac-31a4-4d32-a600-2c274d9535df-kube-api-access-dwvs8\") pod \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\" (UID: \"7f5ec6ac-31a4-4d32-a600-2c274d9535df\") " Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.711826 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f5ec6ac-31a4-4d32-a600-2c274d9535df-kube-api-access-dwvs8" (OuterVolumeSpecName: "kube-api-access-dwvs8") pod "7f5ec6ac-31a4-4d32-a600-2c274d9535df" (UID: "7f5ec6ac-31a4-4d32-a600-2c274d9535df"). InnerVolumeSpecName "kube-api-access-dwvs8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.775680 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwvs8\" (UniqueName: \"kubernetes.io/projected/7f5ec6ac-31a4-4d32-a600-2c274d9535df-kube-api-access-dwvs8\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.833949 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7f5ec6ac-31a4-4d32-a600-2c274d9535df" (UID: "7f5ec6ac-31a4-4d32-a600-2c274d9535df"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.840189 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7f5ec6ac-31a4-4d32-a600-2c274d9535df" (UID: "7f5ec6ac-31a4-4d32-a600-2c274d9535df"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.872936 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7f5ec6ac-31a4-4d32-a600-2c274d9535df" (UID: "7f5ec6ac-31a4-4d32-a600-2c274d9535df"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.874608 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7f5ec6ac-31a4-4d32-a600-2c274d9535df" (UID: "7f5ec6ac-31a4-4d32-a600-2c274d9535df"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.877409 4685 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.877435 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.877444 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.877453 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.909449 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-config" (OuterVolumeSpecName: "config") pod "7f5ec6ac-31a4-4d32-a600-2c274d9535df" (UID: "7f5ec6ac-31a4-4d32-a600-2c274d9535df"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:05:11 crc kubenswrapper[4685]: I1013 09:05:11.978845 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f5ec6ac-31a4-4d32-a600-2c274d9535df-config\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:12 crc kubenswrapper[4685]: I1013 09:05:12.091279 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 09:05:12 crc kubenswrapper[4685]: I1013 09:05:12.126711 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5aa884f8-807f-407b-9ea0-ea0812fcd1a2","Type":"ContainerStarted","Data":"d98a6cff4556381ca5f56f9e11a8cd5a507aad567d44372ae2ee650b381fb3d3"} Oct 13 09:05:12 crc kubenswrapper[4685]: I1013 09:05:12.128803 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" event={"ID":"7f5ec6ac-31a4-4d32-a600-2c274d9535df","Type":"ContainerDied","Data":"2414585a93b2c26acd6f05cc17e9abbf003b285207cc1236089556089138fd27"} Oct 13 09:05:12 crc kubenswrapper[4685]: I1013 09:05:12.128850 4685 scope.go:117] "RemoveContainer" containerID="7ce9dadc75ad7f16a641fb2b4705e647595b8b2a879975af4bb44706eca7143e" Oct 13 09:05:12 crc kubenswrapper[4685]: I1013 09:05:12.128996 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-nv9ck" Oct 13 09:05:12 crc kubenswrapper[4685]: I1013 09:05:12.163524 4685 scope.go:117] "RemoveContainer" containerID="dfa6ba7e634681fda898f0740e271c94b43bc2404d62bfe268cd1ffee6f95b24" Oct 13 09:05:12 crc kubenswrapper[4685]: I1013 09:05:12.184071 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-nv9ck"] Oct 13 09:05:12 crc kubenswrapper[4685]: I1013 09:05:12.189606 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-nv9ck"] Oct 13 09:05:13 crc kubenswrapper[4685]: I1013 09:05:13.147721 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5aa884f8-807f-407b-9ea0-ea0812fcd1a2","Type":"ContainerStarted","Data":"a71816946db98ebb13535ebe86d0c77cdfdea88c0dca9c0f06a70a4295e9ac29"} Oct 13 09:05:13 crc kubenswrapper[4685]: I1013 09:05:13.514771 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f5ec6ac-31a4-4d32-a600-2c274d9535df" path="/var/lib/kubelet/pods/7f5ec6ac-31a4-4d32-a600-2c274d9535df/volumes" Oct 13 09:05:14 crc kubenswrapper[4685]: I1013 09:05:14.160260 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5aa884f8-807f-407b-9ea0-ea0812fcd1a2","Type":"ContainerStarted","Data":"5a504ed428763b84fa468f570db6134807b33fef0aab079a7dc8950c817e107d"} Oct 13 09:05:15 crc kubenswrapper[4685]: I1013 09:05:15.171248 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5aa884f8-807f-407b-9ea0-ea0812fcd1a2","Type":"ContainerStarted","Data":"a7deadfa48737bab94b5e6ed05e83b50c31c02735de3551fa894582bdaa88aff"} Oct 13 09:05:16 crc kubenswrapper[4685]: I1013 09:05:16.182088 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5aa884f8-807f-407b-9ea0-ea0812fcd1a2","Type":"ContainerStarted","Data":"3bb5229111b05ddbb67744a17bbdce17608cf6ec85fcbe9656b571aa0fe8dca8"} Oct 13 09:05:16 crc kubenswrapper[4685]: I1013 09:05:16.183454 4685 generic.go:334] "Generic (PLEG): container finished" podID="fae68cb4-7110-40c0-b48b-b38036250c61" containerID="4387f4f578b34308748b28e4f9426909a73eca7b00aaa8820ff473823a37df64" exitCode=0 Oct 13 09:05:16 crc kubenswrapper[4685]: I1013 09:05:16.183508 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 13 09:05:16 crc kubenswrapper[4685]: I1013 09:05:16.183542 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-tkkqw" event={"ID":"fae68cb4-7110-40c0-b48b-b38036250c61","Type":"ContainerDied","Data":"4387f4f578b34308748b28e4f9426909a73eca7b00aaa8820ff473823a37df64"} Oct 13 09:05:16 crc kubenswrapper[4685]: I1013 09:05:16.218619 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.6225893139999998 podStartE2EDuration="5.218597638s" podCreationTimestamp="2025-10-13 09:05:11 +0000 UTC" firstStartedPulling="2025-10-13 09:05:12.098065534 +0000 UTC m=+1237.245941295" lastFinishedPulling="2025-10-13 09:05:15.694073838 +0000 UTC m=+1240.841949619" observedRunningTime="2025-10-13 09:05:16.201703419 +0000 UTC m=+1241.349579180" watchObservedRunningTime="2025-10-13 09:05:16.218597638 +0000 UTC m=+1241.366473409" Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:17.486869 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:17.490175 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:17.588700 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-tkkqw" Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:17.618479 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fae68cb4-7110-40c0-b48b-b38036250c61-config-data\") pod \"fae68cb4-7110-40c0-b48b-b38036250c61\" (UID: \"fae68cb4-7110-40c0-b48b-b38036250c61\") " Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:17.618558 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fae68cb4-7110-40c0-b48b-b38036250c61-combined-ca-bundle\") pod \"fae68cb4-7110-40c0-b48b-b38036250c61\" (UID: \"fae68cb4-7110-40c0-b48b-b38036250c61\") " Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:17.618593 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fae68cb4-7110-40c0-b48b-b38036250c61-scripts\") pod \"fae68cb4-7110-40c0-b48b-b38036250c61\" (UID: \"fae68cb4-7110-40c0-b48b-b38036250c61\") " Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:17.618636 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvflh\" (UniqueName: \"kubernetes.io/projected/fae68cb4-7110-40c0-b48b-b38036250c61-kube-api-access-jvflh\") pod \"fae68cb4-7110-40c0-b48b-b38036250c61\" (UID: \"fae68cb4-7110-40c0-b48b-b38036250c61\") " Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:17.626584 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fae68cb4-7110-40c0-b48b-b38036250c61-scripts" (OuterVolumeSpecName: "scripts") pod "fae68cb4-7110-40c0-b48b-b38036250c61" (UID: "fae68cb4-7110-40c0-b48b-b38036250c61"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:17.647936 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fae68cb4-7110-40c0-b48b-b38036250c61-kube-api-access-jvflh" (OuterVolumeSpecName: "kube-api-access-jvflh") pod "fae68cb4-7110-40c0-b48b-b38036250c61" (UID: "fae68cb4-7110-40c0-b48b-b38036250c61"). InnerVolumeSpecName "kube-api-access-jvflh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:17.664316 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fae68cb4-7110-40c0-b48b-b38036250c61-config-data" (OuterVolumeSpecName: "config-data") pod "fae68cb4-7110-40c0-b48b-b38036250c61" (UID: "fae68cb4-7110-40c0-b48b-b38036250c61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:17.679433 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fae68cb4-7110-40c0-b48b-b38036250c61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fae68cb4-7110-40c0-b48b-b38036250c61" (UID: "fae68cb4-7110-40c0-b48b-b38036250c61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:17.721138 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fae68cb4-7110-40c0-b48b-b38036250c61-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:17.721167 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fae68cb4-7110-40c0-b48b-b38036250c61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:17.721179 4685 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fae68cb4-7110-40c0-b48b-b38036250c61-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:17.721188 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvflh\" (UniqueName: \"kubernetes.io/projected/fae68cb4-7110-40c0-b48b-b38036250c61-kube-api-access-jvflh\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:18.202690 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-tkkqw" event={"ID":"fae68cb4-7110-40c0-b48b-b38036250c61","Type":"ContainerDied","Data":"868fd71d1609daa2498ee9dcd81aaff3726eb2f2171ce0bb7b20744b20b865e2"} Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:18.202991 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="868fd71d1609daa2498ee9dcd81aaff3726eb2f2171ce0bb7b20744b20b865e2" Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:18.203041 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-tkkqw" Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:18.487463 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:18.487884 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="306676c4-74d1-4de3-8b41-b16ee72504c7" containerName="nova-scheduler-scheduler" containerID="cri-o://e5a829666a35fab0d360b6c0caf2d2ff2502b0a900fe7e7811827b1d7f1e2224" gracePeriod=30 Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:18.502128 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d5931019-0ca3-4483-a465-a7a9fc674179" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:18.502160 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d5931019-0ca3-4483-a465-a7a9fc674179" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:18.563617 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:18.563860 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c2e61fc8-e032-46a2-822d-3bf80fdf77db" containerName="nova-metadata-log" containerID="cri-o://9daae031bca95d2b021c73f80921e5f7417191147f6e574f9d4f4f2fc0c48cc8" gracePeriod=30 Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:18.563960 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c2e61fc8-e032-46a2-822d-3bf80fdf77db" containerName="nova-metadata-metadata" containerID="cri-o://b6c6a2041218515348db8c277deabfd954251d1019139c6e828f0d4b2d0c5ad6" gracePeriod=30 Oct 13 09:05:18 crc kubenswrapper[4685]: I1013 09:05:18.623162 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 13 09:05:19 crc kubenswrapper[4685]: I1013 09:05:19.212178 4685 generic.go:334] "Generic (PLEG): container finished" podID="c2e61fc8-e032-46a2-822d-3bf80fdf77db" containerID="9daae031bca95d2b021c73f80921e5f7417191147f6e574f9d4f4f2fc0c48cc8" exitCode=143 Oct 13 09:05:19 crc kubenswrapper[4685]: I1013 09:05:19.212247 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2e61fc8-e032-46a2-822d-3bf80fdf77db","Type":"ContainerDied","Data":"9daae031bca95d2b021c73f80921e5f7417191147f6e574f9d4f4f2fc0c48cc8"} Oct 13 09:05:19 crc kubenswrapper[4685]: I1013 09:05:19.212365 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d5931019-0ca3-4483-a465-a7a9fc674179" containerName="nova-api-log" containerID="cri-o://d8134cc82308c9452bed5425b7c4dab52375247bb537ed439f0b98fc2de3bf88" gracePeriod=30 Oct 13 09:05:19 crc kubenswrapper[4685]: I1013 09:05:19.212446 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d5931019-0ca3-4483-a465-a7a9fc674179" containerName="nova-api-api" containerID="cri-o://79ea6cf83c728649d57afbd4d598c450785086141be3dc578e47d9f5ea8a6414" gracePeriod=30 Oct 13 09:05:20 crc kubenswrapper[4685]: E1013 09:05:20.009118 4685 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e5a829666a35fab0d360b6c0caf2d2ff2502b0a900fe7e7811827b1d7f1e2224 is running failed: container process not found" containerID="e5a829666a35fab0d360b6c0caf2d2ff2502b0a900fe7e7811827b1d7f1e2224" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 13 09:05:20 crc kubenswrapper[4685]: E1013 09:05:20.009849 4685 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e5a829666a35fab0d360b6c0caf2d2ff2502b0a900fe7e7811827b1d7f1e2224 is running failed: container process not found" containerID="e5a829666a35fab0d360b6c0caf2d2ff2502b0a900fe7e7811827b1d7f1e2224" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 13 09:05:20 crc kubenswrapper[4685]: E1013 09:05:20.012427 4685 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e5a829666a35fab0d360b6c0caf2d2ff2502b0a900fe7e7811827b1d7f1e2224 is running failed: container process not found" containerID="e5a829666a35fab0d360b6c0caf2d2ff2502b0a900fe7e7811827b1d7f1e2224" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 13 09:05:20 crc kubenswrapper[4685]: E1013 09:05:20.012472 4685 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e5a829666a35fab0d360b6c0caf2d2ff2502b0a900fe7e7811827b1d7f1e2224 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="306676c4-74d1-4de3-8b41-b16ee72504c7" containerName="nova-scheduler-scheduler" Oct 13 09:05:20 crc kubenswrapper[4685]: I1013 09:05:20.222484 4685 generic.go:334] "Generic (PLEG): container finished" podID="306676c4-74d1-4de3-8b41-b16ee72504c7" containerID="e5a829666a35fab0d360b6c0caf2d2ff2502b0a900fe7e7811827b1d7f1e2224" exitCode=0 Oct 13 09:05:20 crc kubenswrapper[4685]: I1013 09:05:20.222560 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"306676c4-74d1-4de3-8b41-b16ee72504c7","Type":"ContainerDied","Data":"e5a829666a35fab0d360b6c0caf2d2ff2502b0a900fe7e7811827b1d7f1e2224"} Oct 13 09:05:20 crc kubenswrapper[4685]: I1013 09:05:20.222868 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"306676c4-74d1-4de3-8b41-b16ee72504c7","Type":"ContainerDied","Data":"9d77a8548eb8532cb97e2cbc76a2911548749e7caf9631b68549f63da5781f69"} Oct 13 09:05:20 crc kubenswrapper[4685]: I1013 09:05:20.222882 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d77a8548eb8532cb97e2cbc76a2911548749e7caf9631b68549f63da5781f69" Oct 13 09:05:20 crc kubenswrapper[4685]: I1013 09:05:20.224244 4685 generic.go:334] "Generic (PLEG): container finished" podID="d5931019-0ca3-4483-a465-a7a9fc674179" containerID="d8134cc82308c9452bed5425b7c4dab52375247bb537ed439f0b98fc2de3bf88" exitCode=143 Oct 13 09:05:20 crc kubenswrapper[4685]: I1013 09:05:20.224275 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d5931019-0ca3-4483-a465-a7a9fc674179","Type":"ContainerDied","Data":"d8134cc82308c9452bed5425b7c4dab52375247bb537ed439f0b98fc2de3bf88"} Oct 13 09:05:20 crc kubenswrapper[4685]: I1013 09:05:20.247344 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 13 09:05:20 crc kubenswrapper[4685]: I1013 09:05:20.369583 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/306676c4-74d1-4de3-8b41-b16ee72504c7-combined-ca-bundle\") pod \"306676c4-74d1-4de3-8b41-b16ee72504c7\" (UID: \"306676c4-74d1-4de3-8b41-b16ee72504c7\") " Oct 13 09:05:20 crc kubenswrapper[4685]: I1013 09:05:20.369822 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wr9p\" (UniqueName: \"kubernetes.io/projected/306676c4-74d1-4de3-8b41-b16ee72504c7-kube-api-access-7wr9p\") pod \"306676c4-74d1-4de3-8b41-b16ee72504c7\" (UID: \"306676c4-74d1-4de3-8b41-b16ee72504c7\") " Oct 13 09:05:20 crc kubenswrapper[4685]: I1013 09:05:20.369971 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/306676c4-74d1-4de3-8b41-b16ee72504c7-config-data\") pod \"306676c4-74d1-4de3-8b41-b16ee72504c7\" (UID: \"306676c4-74d1-4de3-8b41-b16ee72504c7\") " Oct 13 09:05:20 crc kubenswrapper[4685]: I1013 09:05:20.378333 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/306676c4-74d1-4de3-8b41-b16ee72504c7-kube-api-access-7wr9p" (OuterVolumeSpecName: "kube-api-access-7wr9p") pod "306676c4-74d1-4de3-8b41-b16ee72504c7" (UID: "306676c4-74d1-4de3-8b41-b16ee72504c7"). InnerVolumeSpecName "kube-api-access-7wr9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:05:20 crc kubenswrapper[4685]: I1013 09:05:20.401905 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/306676c4-74d1-4de3-8b41-b16ee72504c7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "306676c4-74d1-4de3-8b41-b16ee72504c7" (UID: "306676c4-74d1-4de3-8b41-b16ee72504c7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:05:20 crc kubenswrapper[4685]: I1013 09:05:20.424211 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/306676c4-74d1-4de3-8b41-b16ee72504c7-config-data" (OuterVolumeSpecName: "config-data") pod "306676c4-74d1-4de3-8b41-b16ee72504c7" (UID: "306676c4-74d1-4de3-8b41-b16ee72504c7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:05:20 crc kubenswrapper[4685]: I1013 09:05:20.471767 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/306676c4-74d1-4de3-8b41-b16ee72504c7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:20 crc kubenswrapper[4685]: I1013 09:05:20.471799 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wr9p\" (UniqueName: \"kubernetes.io/projected/306676c4-74d1-4de3-8b41-b16ee72504c7-kube-api-access-7wr9p\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:20 crc kubenswrapper[4685]: I1013 09:05:20.471811 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/306676c4-74d1-4de3-8b41-b16ee72504c7-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.231412 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.259173 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.267400 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.280602 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 09:05:21 crc kubenswrapper[4685]: E1013 09:05:21.280981 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f5ec6ac-31a4-4d32-a600-2c274d9535df" containerName="dnsmasq-dns" Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.280994 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f5ec6ac-31a4-4d32-a600-2c274d9535df" containerName="dnsmasq-dns" Oct 13 09:05:21 crc kubenswrapper[4685]: E1013 09:05:21.281024 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fae68cb4-7110-40c0-b48b-b38036250c61" containerName="nova-manage" Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.281030 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="fae68cb4-7110-40c0-b48b-b38036250c61" containerName="nova-manage" Oct 13 09:05:21 crc kubenswrapper[4685]: E1013 09:05:21.281045 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f5ec6ac-31a4-4d32-a600-2c274d9535df" containerName="init" Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.281052 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f5ec6ac-31a4-4d32-a600-2c274d9535df" containerName="init" Oct 13 09:05:21 crc kubenswrapper[4685]: E1013 09:05:21.281062 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="306676c4-74d1-4de3-8b41-b16ee72504c7" containerName="nova-scheduler-scheduler" Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.281068 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="306676c4-74d1-4de3-8b41-b16ee72504c7" containerName="nova-scheduler-scheduler" Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.281249 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="fae68cb4-7110-40c0-b48b-b38036250c61" containerName="nova-manage" Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.281262 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f5ec6ac-31a4-4d32-a600-2c274d9535df" containerName="dnsmasq-dns" Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.281280 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="306676c4-74d1-4de3-8b41-b16ee72504c7" containerName="nova-scheduler-scheduler" Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.281821 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.285512 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c01b8af0-dd0a-4f1d-9e09-1e6ab0cf7779-config-data\") pod \"nova-scheduler-0\" (UID: \"c01b8af0-dd0a-4f1d-9e09-1e6ab0cf7779\") " pod="openstack/nova-scheduler-0" Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.285581 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c01b8af0-dd0a-4f1d-9e09-1e6ab0cf7779-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c01b8af0-dd0a-4f1d-9e09-1e6ab0cf7779\") " pod="openstack/nova-scheduler-0" Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.285603 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf62k\" (UniqueName: \"kubernetes.io/projected/c01b8af0-dd0a-4f1d-9e09-1e6ab0cf7779-kube-api-access-mf62k\") pod \"nova-scheduler-0\" (UID: \"c01b8af0-dd0a-4f1d-9e09-1e6ab0cf7779\") " pod="openstack/nova-scheduler-0" Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.285874 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.306313 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.387289 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c01b8af0-dd0a-4f1d-9e09-1e6ab0cf7779-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c01b8af0-dd0a-4f1d-9e09-1e6ab0cf7779\") " pod="openstack/nova-scheduler-0" Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.387692 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf62k\" (UniqueName: \"kubernetes.io/projected/c01b8af0-dd0a-4f1d-9e09-1e6ab0cf7779-kube-api-access-mf62k\") pod \"nova-scheduler-0\" (UID: \"c01b8af0-dd0a-4f1d-9e09-1e6ab0cf7779\") " pod="openstack/nova-scheduler-0" Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.387840 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c01b8af0-dd0a-4f1d-9e09-1e6ab0cf7779-config-data\") pod \"nova-scheduler-0\" (UID: \"c01b8af0-dd0a-4f1d-9e09-1e6ab0cf7779\") " pod="openstack/nova-scheduler-0" Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.391443 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c01b8af0-dd0a-4f1d-9e09-1e6ab0cf7779-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c01b8af0-dd0a-4f1d-9e09-1e6ab0cf7779\") " pod="openstack/nova-scheduler-0" Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.395118 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c01b8af0-dd0a-4f1d-9e09-1e6ab0cf7779-config-data\") pod \"nova-scheduler-0\" (UID: \"c01b8af0-dd0a-4f1d-9e09-1e6ab0cf7779\") " pod="openstack/nova-scheduler-0" Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.404144 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf62k\" (UniqueName: \"kubernetes.io/projected/c01b8af0-dd0a-4f1d-9e09-1e6ab0cf7779-kube-api-access-mf62k\") pod \"nova-scheduler-0\" (UID: \"c01b8af0-dd0a-4f1d-9e09-1e6ab0cf7779\") " pod="openstack/nova-scheduler-0" Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.520601 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="306676c4-74d1-4de3-8b41-b16ee72504c7" path="/var/lib/kubelet/pods/306676c4-74d1-4de3-8b41-b16ee72504c7/volumes" Oct 13 09:05:21 crc kubenswrapper[4685]: I1013 09:05:21.623947 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 13 09:05:22 crc kubenswrapper[4685]: I1013 09:05:22.110359 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 09:05:22 crc kubenswrapper[4685]: W1013 09:05:22.114153 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc01b8af0_dd0a_4f1d_9e09_1e6ab0cf7779.slice/crio-daf0ea09fbca973363913041d0acaee5bde39849c41cd7e27396115063e45e1f WatchSource:0}: Error finding container daf0ea09fbca973363913041d0acaee5bde39849c41cd7e27396115063e45e1f: Status 404 returned error can't find the container with id daf0ea09fbca973363913041d0acaee5bde39849c41cd7e27396115063e45e1f Oct 13 09:05:22 crc kubenswrapper[4685]: I1013 09:05:22.256068 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c01b8af0-dd0a-4f1d-9e09-1e6ab0cf7779","Type":"ContainerStarted","Data":"daf0ea09fbca973363913041d0acaee5bde39849c41cd7e27396115063e45e1f"} Oct 13 09:05:22 crc kubenswrapper[4685]: I1013 09:05:22.682383 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 09:05:22 crc kubenswrapper[4685]: I1013 09:05:22.817469 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7klzs\" (UniqueName: \"kubernetes.io/projected/c2e61fc8-e032-46a2-822d-3bf80fdf77db-kube-api-access-7klzs\") pod \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\" (UID: \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\") " Oct 13 09:05:22 crc kubenswrapper[4685]: I1013 09:05:22.817560 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2e61fc8-e032-46a2-822d-3bf80fdf77db-config-data\") pod \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\" (UID: \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\") " Oct 13 09:05:22 crc kubenswrapper[4685]: I1013 09:05:22.817590 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e61fc8-e032-46a2-822d-3bf80fdf77db-combined-ca-bundle\") pod \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\" (UID: \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\") " Oct 13 09:05:22 crc kubenswrapper[4685]: I1013 09:05:22.817621 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2e61fc8-e032-46a2-822d-3bf80fdf77db-logs\") pod \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\" (UID: \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\") " Oct 13 09:05:22 crc kubenswrapper[4685]: I1013 09:05:22.817662 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2e61fc8-e032-46a2-822d-3bf80fdf77db-nova-metadata-tls-certs\") pod \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\" (UID: \"c2e61fc8-e032-46a2-822d-3bf80fdf77db\") " Oct 13 09:05:22 crc kubenswrapper[4685]: I1013 09:05:22.818543 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2e61fc8-e032-46a2-822d-3bf80fdf77db-logs" (OuterVolumeSpecName: "logs") pod "c2e61fc8-e032-46a2-822d-3bf80fdf77db" (UID: "c2e61fc8-e032-46a2-822d-3bf80fdf77db"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:05:22 crc kubenswrapper[4685]: I1013 09:05:22.840238 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2e61fc8-e032-46a2-822d-3bf80fdf77db-kube-api-access-7klzs" (OuterVolumeSpecName: "kube-api-access-7klzs") pod "c2e61fc8-e032-46a2-822d-3bf80fdf77db" (UID: "c2e61fc8-e032-46a2-822d-3bf80fdf77db"). InnerVolumeSpecName "kube-api-access-7klzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:05:22 crc kubenswrapper[4685]: I1013 09:05:22.845856 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e61fc8-e032-46a2-822d-3bf80fdf77db-config-data" (OuterVolumeSpecName: "config-data") pod "c2e61fc8-e032-46a2-822d-3bf80fdf77db" (UID: "c2e61fc8-e032-46a2-822d-3bf80fdf77db"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:05:22 crc kubenswrapper[4685]: I1013 09:05:22.872131 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e61fc8-e032-46a2-822d-3bf80fdf77db-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c2e61fc8-e032-46a2-822d-3bf80fdf77db" (UID: "c2e61fc8-e032-46a2-822d-3bf80fdf77db"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:05:22 crc kubenswrapper[4685]: I1013 09:05:22.874378 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e61fc8-e032-46a2-822d-3bf80fdf77db-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "c2e61fc8-e032-46a2-822d-3bf80fdf77db" (UID: "c2e61fc8-e032-46a2-822d-3bf80fdf77db"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:05:22 crc kubenswrapper[4685]: I1013 09:05:22.919364 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2e61fc8-e032-46a2-822d-3bf80fdf77db-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:22 crc kubenswrapper[4685]: I1013 09:05:22.919399 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e61fc8-e032-46a2-822d-3bf80fdf77db-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:22 crc kubenswrapper[4685]: I1013 09:05:22.919414 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2e61fc8-e032-46a2-822d-3bf80fdf77db-logs\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:22 crc kubenswrapper[4685]: I1013 09:05:22.919425 4685 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2e61fc8-e032-46a2-822d-3bf80fdf77db-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:22 crc kubenswrapper[4685]: I1013 09:05:22.919435 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7klzs\" (UniqueName: \"kubernetes.io/projected/c2e61fc8-e032-46a2-822d-3bf80fdf77db-kube-api-access-7klzs\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:22 crc kubenswrapper[4685]: I1013 09:05:22.980294 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:05:22 crc kubenswrapper[4685]: I1013 09:05:22.980357 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.265331 4685 generic.go:334] "Generic (PLEG): container finished" podID="c2e61fc8-e032-46a2-822d-3bf80fdf77db" containerID="b6c6a2041218515348db8c277deabfd954251d1019139c6e828f0d4b2d0c5ad6" exitCode=0 Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.265373 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2e61fc8-e032-46a2-822d-3bf80fdf77db","Type":"ContainerDied","Data":"b6c6a2041218515348db8c277deabfd954251d1019139c6e828f0d4b2d0c5ad6"} Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.265393 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.265637 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2e61fc8-e032-46a2-822d-3bf80fdf77db","Type":"ContainerDied","Data":"e315ca5134fa7db328cde936ab3c9e03300249c8564811529626feab2f12297f"} Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.265668 4685 scope.go:117] "RemoveContainer" containerID="b6c6a2041218515348db8c277deabfd954251d1019139c6e828f0d4b2d0c5ad6" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.268214 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c01b8af0-dd0a-4f1d-9e09-1e6ab0cf7779","Type":"ContainerStarted","Data":"362e86338603cf85f7159b68b906ff7dd85e37c5037f511283598af216261735"} Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.292797 4685 scope.go:117] "RemoveContainer" containerID="9daae031bca95d2b021c73f80921e5f7417191147f6e574f9d4f4f2fc0c48cc8" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.321656 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.321635579 podStartE2EDuration="2.321635579s" podCreationTimestamp="2025-10-13 09:05:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:05:23.300615948 +0000 UTC m=+1248.448491709" watchObservedRunningTime="2025-10-13 09:05:23.321635579 +0000 UTC m=+1248.469511340" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.322561 4685 scope.go:117] "RemoveContainer" containerID="b6c6a2041218515348db8c277deabfd954251d1019139c6e828f0d4b2d0c5ad6" Oct 13 09:05:23 crc kubenswrapper[4685]: E1013 09:05:23.322996 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6c6a2041218515348db8c277deabfd954251d1019139c6e828f0d4b2d0c5ad6\": container with ID starting with b6c6a2041218515348db8c277deabfd954251d1019139c6e828f0d4b2d0c5ad6 not found: ID does not exist" containerID="b6c6a2041218515348db8c277deabfd954251d1019139c6e828f0d4b2d0c5ad6" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.323117 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6c6a2041218515348db8c277deabfd954251d1019139c6e828f0d4b2d0c5ad6"} err="failed to get container status \"b6c6a2041218515348db8c277deabfd954251d1019139c6e828f0d4b2d0c5ad6\": rpc error: code = NotFound desc = could not find container \"b6c6a2041218515348db8c277deabfd954251d1019139c6e828f0d4b2d0c5ad6\": container with ID starting with b6c6a2041218515348db8c277deabfd954251d1019139c6e828f0d4b2d0c5ad6 not found: ID does not exist" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.323228 4685 scope.go:117] "RemoveContainer" containerID="9daae031bca95d2b021c73f80921e5f7417191147f6e574f9d4f4f2fc0c48cc8" Oct 13 09:05:23 crc kubenswrapper[4685]: E1013 09:05:23.323525 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9daae031bca95d2b021c73f80921e5f7417191147f6e574f9d4f4f2fc0c48cc8\": container with ID starting with 9daae031bca95d2b021c73f80921e5f7417191147f6e574f9d4f4f2fc0c48cc8 not found: ID does not exist" containerID="9daae031bca95d2b021c73f80921e5f7417191147f6e574f9d4f4f2fc0c48cc8" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.323628 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9daae031bca95d2b021c73f80921e5f7417191147f6e574f9d4f4f2fc0c48cc8"} err="failed to get container status \"9daae031bca95d2b021c73f80921e5f7417191147f6e574f9d4f4f2fc0c48cc8\": rpc error: code = NotFound desc = could not find container \"9daae031bca95d2b021c73f80921e5f7417191147f6e574f9d4f4f2fc0c48cc8\": container with ID starting with 9daae031bca95d2b021c73f80921e5f7417191147f6e574f9d4f4f2fc0c48cc8 not found: ID does not exist" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.337070 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.345116 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.356252 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 13 09:05:23 crc kubenswrapper[4685]: E1013 09:05:23.357011 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2e61fc8-e032-46a2-822d-3bf80fdf77db" containerName="nova-metadata-metadata" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.357024 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2e61fc8-e032-46a2-822d-3bf80fdf77db" containerName="nova-metadata-metadata" Oct 13 09:05:23 crc kubenswrapper[4685]: E1013 09:05:23.357062 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2e61fc8-e032-46a2-822d-3bf80fdf77db" containerName="nova-metadata-log" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.357069 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2e61fc8-e032-46a2-822d-3bf80fdf77db" containerName="nova-metadata-log" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.357443 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2e61fc8-e032-46a2-822d-3bf80fdf77db" containerName="nova-metadata-log" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.357482 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2e61fc8-e032-46a2-822d-3bf80fdf77db" containerName="nova-metadata-metadata" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.363524 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.368656 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.368928 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.375459 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.431560 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftftk\" (UniqueName: \"kubernetes.io/projected/7ddd9484-88bb-4550-b9e7-399e001cfa29-kube-api-access-ftftk\") pod \"nova-metadata-0\" (UID: \"7ddd9484-88bb-4550-b9e7-399e001cfa29\") " pod="openstack/nova-metadata-0" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.431637 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ddd9484-88bb-4550-b9e7-399e001cfa29-logs\") pod \"nova-metadata-0\" (UID: \"7ddd9484-88bb-4550-b9e7-399e001cfa29\") " pod="openstack/nova-metadata-0" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.431756 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ddd9484-88bb-4550-b9e7-399e001cfa29-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7ddd9484-88bb-4550-b9e7-399e001cfa29\") " pod="openstack/nova-metadata-0" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.431812 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ddd9484-88bb-4550-b9e7-399e001cfa29-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7ddd9484-88bb-4550-b9e7-399e001cfa29\") " pod="openstack/nova-metadata-0" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.431842 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ddd9484-88bb-4550-b9e7-399e001cfa29-config-data\") pod \"nova-metadata-0\" (UID: \"7ddd9484-88bb-4550-b9e7-399e001cfa29\") " pod="openstack/nova-metadata-0" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.516281 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2e61fc8-e032-46a2-822d-3bf80fdf77db" path="/var/lib/kubelet/pods/c2e61fc8-e032-46a2-822d-3bf80fdf77db/volumes" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.533561 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftftk\" (UniqueName: \"kubernetes.io/projected/7ddd9484-88bb-4550-b9e7-399e001cfa29-kube-api-access-ftftk\") pod \"nova-metadata-0\" (UID: \"7ddd9484-88bb-4550-b9e7-399e001cfa29\") " pod="openstack/nova-metadata-0" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.533611 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ddd9484-88bb-4550-b9e7-399e001cfa29-logs\") pod \"nova-metadata-0\" (UID: \"7ddd9484-88bb-4550-b9e7-399e001cfa29\") " pod="openstack/nova-metadata-0" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.533700 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ddd9484-88bb-4550-b9e7-399e001cfa29-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7ddd9484-88bb-4550-b9e7-399e001cfa29\") " pod="openstack/nova-metadata-0" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.533745 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ddd9484-88bb-4550-b9e7-399e001cfa29-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7ddd9484-88bb-4550-b9e7-399e001cfa29\") " pod="openstack/nova-metadata-0" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.533771 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ddd9484-88bb-4550-b9e7-399e001cfa29-config-data\") pod \"nova-metadata-0\" (UID: \"7ddd9484-88bb-4550-b9e7-399e001cfa29\") " pod="openstack/nova-metadata-0" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.534077 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ddd9484-88bb-4550-b9e7-399e001cfa29-logs\") pod \"nova-metadata-0\" (UID: \"7ddd9484-88bb-4550-b9e7-399e001cfa29\") " pod="openstack/nova-metadata-0" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.538266 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ddd9484-88bb-4550-b9e7-399e001cfa29-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7ddd9484-88bb-4550-b9e7-399e001cfa29\") " pod="openstack/nova-metadata-0" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.539301 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ddd9484-88bb-4550-b9e7-399e001cfa29-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7ddd9484-88bb-4550-b9e7-399e001cfa29\") " pod="openstack/nova-metadata-0" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.544299 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ddd9484-88bb-4550-b9e7-399e001cfa29-config-data\") pod \"nova-metadata-0\" (UID: \"7ddd9484-88bb-4550-b9e7-399e001cfa29\") " pod="openstack/nova-metadata-0" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.560558 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftftk\" (UniqueName: \"kubernetes.io/projected/7ddd9484-88bb-4550-b9e7-399e001cfa29-kube-api-access-ftftk\") pod \"nova-metadata-0\" (UID: \"7ddd9484-88bb-4550-b9e7-399e001cfa29\") " pod="openstack/nova-metadata-0" Oct 13 09:05:23 crc kubenswrapper[4685]: I1013 09:05:23.696568 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 09:05:24 crc kubenswrapper[4685]: W1013 09:05:24.219025 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ddd9484_88bb_4550_b9e7_399e001cfa29.slice/crio-566472fbdec9e06043009f9b2d3711554212fa8205a438b3f42147fe131b55bd WatchSource:0}: Error finding container 566472fbdec9e06043009f9b2d3711554212fa8205a438b3f42147fe131b55bd: Status 404 returned error can't find the container with id 566472fbdec9e06043009f9b2d3711554212fa8205a438b3f42147fe131b55bd Oct 13 09:05:24 crc kubenswrapper[4685]: I1013 09:05:24.227499 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 09:05:24 crc kubenswrapper[4685]: I1013 09:05:24.281609 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7ddd9484-88bb-4550-b9e7-399e001cfa29","Type":"ContainerStarted","Data":"566472fbdec9e06043009f9b2d3711554212fa8205a438b3f42147fe131b55bd"} Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.021073 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.061296 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-internal-tls-certs\") pod \"d5931019-0ca3-4483-a465-a7a9fc674179\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.061365 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5931019-0ca3-4483-a465-a7a9fc674179-logs\") pod \"d5931019-0ca3-4483-a465-a7a9fc674179\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.061408 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-combined-ca-bundle\") pod \"d5931019-0ca3-4483-a465-a7a9fc674179\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.061447 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-public-tls-certs\") pod \"d5931019-0ca3-4483-a465-a7a9fc674179\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.061464 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-config-data\") pod \"d5931019-0ca3-4483-a465-a7a9fc674179\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.061507 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25j9c\" (UniqueName: \"kubernetes.io/projected/d5931019-0ca3-4483-a465-a7a9fc674179-kube-api-access-25j9c\") pod \"d5931019-0ca3-4483-a465-a7a9fc674179\" (UID: \"d5931019-0ca3-4483-a465-a7a9fc674179\") " Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.062820 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5931019-0ca3-4483-a465-a7a9fc674179-logs" (OuterVolumeSpecName: "logs") pod "d5931019-0ca3-4483-a465-a7a9fc674179" (UID: "d5931019-0ca3-4483-a465-a7a9fc674179"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.075258 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5931019-0ca3-4483-a465-a7a9fc674179-kube-api-access-25j9c" (OuterVolumeSpecName: "kube-api-access-25j9c") pod "d5931019-0ca3-4483-a465-a7a9fc674179" (UID: "d5931019-0ca3-4483-a465-a7a9fc674179"). InnerVolumeSpecName "kube-api-access-25j9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.110674 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-config-data" (OuterVolumeSpecName: "config-data") pod "d5931019-0ca3-4483-a465-a7a9fc674179" (UID: "d5931019-0ca3-4483-a465-a7a9fc674179"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.112258 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d5931019-0ca3-4483-a465-a7a9fc674179" (UID: "d5931019-0ca3-4483-a465-a7a9fc674179"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.139415 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d5931019-0ca3-4483-a465-a7a9fc674179" (UID: "d5931019-0ca3-4483-a465-a7a9fc674179"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.141329 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d5931019-0ca3-4483-a465-a7a9fc674179" (UID: "d5931019-0ca3-4483-a465-a7a9fc674179"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.164176 4685 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.164206 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.164215 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25j9c\" (UniqueName: \"kubernetes.io/projected/d5931019-0ca3-4483-a465-a7a9fc674179-kube-api-access-25j9c\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.164226 4685 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.164253 4685 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5931019-0ca3-4483-a465-a7a9fc674179-logs\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.164263 4685 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5931019-0ca3-4483-a465-a7a9fc674179-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.307352 4685 generic.go:334] "Generic (PLEG): container finished" podID="d5931019-0ca3-4483-a465-a7a9fc674179" containerID="79ea6cf83c728649d57afbd4d598c450785086141be3dc578e47d9f5ea8a6414" exitCode=0 Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.307493 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.308174 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d5931019-0ca3-4483-a465-a7a9fc674179","Type":"ContainerDied","Data":"79ea6cf83c728649d57afbd4d598c450785086141be3dc578e47d9f5ea8a6414"} Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.308227 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d5931019-0ca3-4483-a465-a7a9fc674179","Type":"ContainerDied","Data":"dfd8b98c6008d550f82ab5ce3828fc6da536fa3e6a5a6c89cf856e6e36441aae"} Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.308246 4685 scope.go:117] "RemoveContainer" containerID="79ea6cf83c728649d57afbd4d598c450785086141be3dc578e47d9f5ea8a6414" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.312658 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7ddd9484-88bb-4550-b9e7-399e001cfa29","Type":"ContainerStarted","Data":"750dd065aed0b5ac3d07935d109936a57483c0246757406509f718a5caf4f916"} Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.312700 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7ddd9484-88bb-4550-b9e7-399e001cfa29","Type":"ContainerStarted","Data":"f51979530cf25351fc3f1716ee2435f2c522cc16be29569363cf08256ace269a"} Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.335553 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.335532866 podStartE2EDuration="2.335532866s" podCreationTimestamp="2025-10-13 09:05:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:05:25.328635279 +0000 UTC m=+1250.476511040" watchObservedRunningTime="2025-10-13 09:05:25.335532866 +0000 UTC m=+1250.483408647" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.350965 4685 scope.go:117] "RemoveContainer" containerID="d8134cc82308c9452bed5425b7c4dab52375247bb537ed439f0b98fc2de3bf88" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.367921 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.377439 4685 scope.go:117] "RemoveContainer" containerID="79ea6cf83c728649d57afbd4d598c450785086141be3dc578e47d9f5ea8a6414" Oct 13 09:05:25 crc kubenswrapper[4685]: E1013 09:05:25.379325 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79ea6cf83c728649d57afbd4d598c450785086141be3dc578e47d9f5ea8a6414\": container with ID starting with 79ea6cf83c728649d57afbd4d598c450785086141be3dc578e47d9f5ea8a6414 not found: ID does not exist" containerID="79ea6cf83c728649d57afbd4d598c450785086141be3dc578e47d9f5ea8a6414" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.379480 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79ea6cf83c728649d57afbd4d598c450785086141be3dc578e47d9f5ea8a6414"} err="failed to get container status \"79ea6cf83c728649d57afbd4d598c450785086141be3dc578e47d9f5ea8a6414\": rpc error: code = NotFound desc = could not find container \"79ea6cf83c728649d57afbd4d598c450785086141be3dc578e47d9f5ea8a6414\": container with ID starting with 79ea6cf83c728649d57afbd4d598c450785086141be3dc578e47d9f5ea8a6414 not found: ID does not exist" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.379584 4685 scope.go:117] "RemoveContainer" containerID="d8134cc82308c9452bed5425b7c4dab52375247bb537ed439f0b98fc2de3bf88" Oct 13 09:05:25 crc kubenswrapper[4685]: E1013 09:05:25.379984 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8134cc82308c9452bed5425b7c4dab52375247bb537ed439f0b98fc2de3bf88\": container with ID starting with d8134cc82308c9452bed5425b7c4dab52375247bb537ed439f0b98fc2de3bf88 not found: ID does not exist" containerID="d8134cc82308c9452bed5425b7c4dab52375247bb537ed439f0b98fc2de3bf88" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.380018 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8134cc82308c9452bed5425b7c4dab52375247bb537ed439f0b98fc2de3bf88"} err="failed to get container status \"d8134cc82308c9452bed5425b7c4dab52375247bb537ed439f0b98fc2de3bf88\": rpc error: code = NotFound desc = could not find container \"d8134cc82308c9452bed5425b7c4dab52375247bb537ed439f0b98fc2de3bf88\": container with ID starting with d8134cc82308c9452bed5425b7c4dab52375247bb537ed439f0b98fc2de3bf88 not found: ID does not exist" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.383712 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.413028 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 13 09:05:25 crc kubenswrapper[4685]: E1013 09:05:25.413703 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5931019-0ca3-4483-a465-a7a9fc674179" containerName="nova-api-log" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.413725 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5931019-0ca3-4483-a465-a7a9fc674179" containerName="nova-api-log" Oct 13 09:05:25 crc kubenswrapper[4685]: E1013 09:05:25.413742 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5931019-0ca3-4483-a465-a7a9fc674179" containerName="nova-api-api" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.413748 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5931019-0ca3-4483-a465-a7a9fc674179" containerName="nova-api-api" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.413982 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5931019-0ca3-4483-a465-a7a9fc674179" containerName="nova-api-api" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.414006 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5931019-0ca3-4483-a465-a7a9fc674179" containerName="nova-api-log" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.415204 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.417272 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.419439 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.421361 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.423069 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.520510 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5931019-0ca3-4483-a465-a7a9fc674179" path="/var/lib/kubelet/pods/d5931019-0ca3-4483-a465-a7a9fc674179/volumes" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.571284 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27a5deb0-f57d-4780-9426-7ff101dc91ed-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"27a5deb0-f57d-4780-9426-7ff101dc91ed\") " pod="openstack/nova-api-0" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.571344 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a5deb0-f57d-4780-9426-7ff101dc91ed-internal-tls-certs\") pod \"nova-api-0\" (UID: \"27a5deb0-f57d-4780-9426-7ff101dc91ed\") " pod="openstack/nova-api-0" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.572390 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27a5deb0-f57d-4780-9426-7ff101dc91ed-logs\") pod \"nova-api-0\" (UID: \"27a5deb0-f57d-4780-9426-7ff101dc91ed\") " pod="openstack/nova-api-0" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.572554 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq4x9\" (UniqueName: \"kubernetes.io/projected/27a5deb0-f57d-4780-9426-7ff101dc91ed-kube-api-access-lq4x9\") pod \"nova-api-0\" (UID: \"27a5deb0-f57d-4780-9426-7ff101dc91ed\") " pod="openstack/nova-api-0" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.572711 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a5deb0-f57d-4780-9426-7ff101dc91ed-public-tls-certs\") pod \"nova-api-0\" (UID: \"27a5deb0-f57d-4780-9426-7ff101dc91ed\") " pod="openstack/nova-api-0" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.572826 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27a5deb0-f57d-4780-9426-7ff101dc91ed-config-data\") pod \"nova-api-0\" (UID: \"27a5deb0-f57d-4780-9426-7ff101dc91ed\") " pod="openstack/nova-api-0" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.674233 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a5deb0-f57d-4780-9426-7ff101dc91ed-public-tls-certs\") pod \"nova-api-0\" (UID: \"27a5deb0-f57d-4780-9426-7ff101dc91ed\") " pod="openstack/nova-api-0" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.674320 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27a5deb0-f57d-4780-9426-7ff101dc91ed-config-data\") pod \"nova-api-0\" (UID: \"27a5deb0-f57d-4780-9426-7ff101dc91ed\") " pod="openstack/nova-api-0" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.674411 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27a5deb0-f57d-4780-9426-7ff101dc91ed-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"27a5deb0-f57d-4780-9426-7ff101dc91ed\") " pod="openstack/nova-api-0" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.674464 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a5deb0-f57d-4780-9426-7ff101dc91ed-internal-tls-certs\") pod \"nova-api-0\" (UID: \"27a5deb0-f57d-4780-9426-7ff101dc91ed\") " pod="openstack/nova-api-0" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.674667 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27a5deb0-f57d-4780-9426-7ff101dc91ed-logs\") pod \"nova-api-0\" (UID: \"27a5deb0-f57d-4780-9426-7ff101dc91ed\") " pod="openstack/nova-api-0" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.675054 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27a5deb0-f57d-4780-9426-7ff101dc91ed-logs\") pod \"nova-api-0\" (UID: \"27a5deb0-f57d-4780-9426-7ff101dc91ed\") " pod="openstack/nova-api-0" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.675249 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq4x9\" (UniqueName: \"kubernetes.io/projected/27a5deb0-f57d-4780-9426-7ff101dc91ed-kube-api-access-lq4x9\") pod \"nova-api-0\" (UID: \"27a5deb0-f57d-4780-9426-7ff101dc91ed\") " pod="openstack/nova-api-0" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.681130 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27a5deb0-f57d-4780-9426-7ff101dc91ed-config-data\") pod \"nova-api-0\" (UID: \"27a5deb0-f57d-4780-9426-7ff101dc91ed\") " pod="openstack/nova-api-0" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.682491 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a5deb0-f57d-4780-9426-7ff101dc91ed-public-tls-certs\") pod \"nova-api-0\" (UID: \"27a5deb0-f57d-4780-9426-7ff101dc91ed\") " pod="openstack/nova-api-0" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.683663 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27a5deb0-f57d-4780-9426-7ff101dc91ed-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"27a5deb0-f57d-4780-9426-7ff101dc91ed\") " pod="openstack/nova-api-0" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.690627 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27a5deb0-f57d-4780-9426-7ff101dc91ed-internal-tls-certs\") pod \"nova-api-0\" (UID: \"27a5deb0-f57d-4780-9426-7ff101dc91ed\") " pod="openstack/nova-api-0" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.691448 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq4x9\" (UniqueName: \"kubernetes.io/projected/27a5deb0-f57d-4780-9426-7ff101dc91ed-kube-api-access-lq4x9\") pod \"nova-api-0\" (UID: \"27a5deb0-f57d-4780-9426-7ff101dc91ed\") " pod="openstack/nova-api-0" Oct 13 09:05:25 crc kubenswrapper[4685]: I1013 09:05:25.756246 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 09:05:26 crc kubenswrapper[4685]: W1013 09:05:26.193754 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27a5deb0_f57d_4780_9426_7ff101dc91ed.slice/crio-2fe653e416a355dcb9629b076d0a3c0f7d1f46d9c4d40f2b2fe668801cc54d2a WatchSource:0}: Error finding container 2fe653e416a355dcb9629b076d0a3c0f7d1f46d9c4d40f2b2fe668801cc54d2a: Status 404 returned error can't find the container with id 2fe653e416a355dcb9629b076d0a3c0f7d1f46d9c4d40f2b2fe668801cc54d2a Oct 13 09:05:26 crc kubenswrapper[4685]: I1013 09:05:26.196949 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 13 09:05:26 crc kubenswrapper[4685]: I1013 09:05:26.329199 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"27a5deb0-f57d-4780-9426-7ff101dc91ed","Type":"ContainerStarted","Data":"2fe653e416a355dcb9629b076d0a3c0f7d1f46d9c4d40f2b2fe668801cc54d2a"} Oct 13 09:05:26 crc kubenswrapper[4685]: I1013 09:05:26.624204 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 13 09:05:27 crc kubenswrapper[4685]: I1013 09:05:27.342415 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"27a5deb0-f57d-4780-9426-7ff101dc91ed","Type":"ContainerStarted","Data":"085d038450a8c4ed0bae8b27eb28cfee754069a7064b8010bda4312f731c7b00"} Oct 13 09:05:27 crc kubenswrapper[4685]: I1013 09:05:27.342462 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"27a5deb0-f57d-4780-9426-7ff101dc91ed","Type":"ContainerStarted","Data":"469dbc12cabcc18eb9fa399816f4a68e373fdb2eff90576efa4da3da9cb35c20"} Oct 13 09:05:27 crc kubenswrapper[4685]: I1013 09:05:27.362608 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.362592451 podStartE2EDuration="2.362592451s" podCreationTimestamp="2025-10-13 09:05:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:05:27.360020841 +0000 UTC m=+1252.507896642" watchObservedRunningTime="2025-10-13 09:05:27.362592451 +0000 UTC m=+1252.510468212" Oct 13 09:05:28 crc kubenswrapper[4685]: I1013 09:05:28.697233 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 13 09:05:28 crc kubenswrapper[4685]: I1013 09:05:28.697563 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 13 09:05:31 crc kubenswrapper[4685]: I1013 09:05:31.624080 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 13 09:05:31 crc kubenswrapper[4685]: I1013 09:05:31.661300 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 13 09:05:32 crc kubenswrapper[4685]: I1013 09:05:32.443774 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 13 09:05:33 crc kubenswrapper[4685]: I1013 09:05:33.696706 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 13 09:05:33 crc kubenswrapper[4685]: I1013 09:05:33.697031 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 13 09:05:34 crc kubenswrapper[4685]: I1013 09:05:34.710134 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7ddd9484-88bb-4550-b9e7-399e001cfa29" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.208:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 13 09:05:34 crc kubenswrapper[4685]: I1013 09:05:34.710146 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7ddd9484-88bb-4550-b9e7-399e001cfa29" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.208:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 13 09:05:35 crc kubenswrapper[4685]: I1013 09:05:35.757045 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 13 09:05:35 crc kubenswrapper[4685]: I1013 09:05:35.757294 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 13 09:05:36 crc kubenswrapper[4685]: I1013 09:05:36.767151 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="27a5deb0-f57d-4780-9426-7ff101dc91ed" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.209:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 13 09:05:36 crc kubenswrapper[4685]: I1013 09:05:36.767180 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="27a5deb0-f57d-4780-9426-7ff101dc91ed" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.209:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 13 09:05:41 crc kubenswrapper[4685]: I1013 09:05:41.715556 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 13 09:05:43 crc kubenswrapper[4685]: I1013 09:05:43.701766 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 13 09:05:43 crc kubenswrapper[4685]: I1013 09:05:43.703787 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 13 09:05:43 crc kubenswrapper[4685]: I1013 09:05:43.707191 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 13 09:05:44 crc kubenswrapper[4685]: I1013 09:05:44.547123 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 13 09:05:45 crc kubenswrapper[4685]: I1013 09:05:45.765148 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 13 09:05:45 crc kubenswrapper[4685]: I1013 09:05:45.766396 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 13 09:05:45 crc kubenswrapper[4685]: I1013 09:05:45.771925 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 13 09:05:45 crc kubenswrapper[4685]: I1013 09:05:45.792680 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 13 09:05:46 crc kubenswrapper[4685]: I1013 09:05:46.551599 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 13 09:05:46 crc kubenswrapper[4685]: I1013 09:05:46.557371 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 13 09:05:52 crc kubenswrapper[4685]: I1013 09:05:52.980737 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:05:52 crc kubenswrapper[4685]: I1013 09:05:52.981611 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:05:52 crc kubenswrapper[4685]: I1013 09:05:52.981692 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 09:05:52 crc kubenswrapper[4685]: I1013 09:05:52.982979 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3648096a0845e8e530c524922c94cf62c1c3b816bc3674fd046d4900ce16a389"} pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 09:05:52 crc kubenswrapper[4685]: I1013 09:05:52.983094 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" containerID="cri-o://3648096a0845e8e530c524922c94cf62c1c3b816bc3674fd046d4900ce16a389" gracePeriod=600 Oct 13 09:05:53 crc kubenswrapper[4685]: I1013 09:05:53.669540 4685 generic.go:334] "Generic (PLEG): container finished" podID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerID="3648096a0845e8e530c524922c94cf62c1c3b816bc3674fd046d4900ce16a389" exitCode=0 Oct 13 09:05:53 crc kubenswrapper[4685]: I1013 09:05:53.669592 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerDied","Data":"3648096a0845e8e530c524922c94cf62c1c3b816bc3674fd046d4900ce16a389"} Oct 13 09:05:53 crc kubenswrapper[4685]: I1013 09:05:53.669903 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerStarted","Data":"a2277b2f3c48d695283091e6c5552e7453d3c307f3bc3a11a7a1f02c1122c7c2"} Oct 13 09:05:53 crc kubenswrapper[4685]: I1013 09:05:53.669946 4685 scope.go:117] "RemoveContainer" containerID="722c93d4dbc8e9002b13fe5143e24de1b4add33c78bcd87a264b498be58dd6fd" Oct 13 09:05:55 crc kubenswrapper[4685]: I1013 09:05:55.287792 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 13 09:05:56 crc kubenswrapper[4685]: I1013 09:05:56.138444 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 13 09:06:00 crc kubenswrapper[4685]: I1013 09:06:00.298322 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="3b0348a6-6814-47f3-919e-e975c01e00e8" containerName="rabbitmq" containerID="cri-o://692441bc08af17df5ba27504a2d82f321da33197cff11964acc63b131225301e" gracePeriod=604795 Oct 13 09:06:00 crc kubenswrapper[4685]: I1013 09:06:00.735437 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="b7f19239-a2d9-49b5-80b0-43543daf1787" containerName="rabbitmq" containerID="cri-o://0ab5c1c1eaab4696cc58a8975e88c605263b37c6525d8d978b6c52d12ae3182e" gracePeriod=604796 Oct 13 09:06:04 crc kubenswrapper[4685]: I1013 09:06:04.938300 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="3b0348a6-6814-47f3-919e-e975c01e00e8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Oct 13 09:06:05 crc kubenswrapper[4685]: I1013 09:06:05.005002 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="b7f19239-a2d9-49b5-80b0-43543daf1787" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Oct 13 09:06:06 crc kubenswrapper[4685]: I1013 09:06:06.794771 4685 generic.go:334] "Generic (PLEG): container finished" podID="3b0348a6-6814-47f3-919e-e975c01e00e8" containerID="692441bc08af17df5ba27504a2d82f321da33197cff11964acc63b131225301e" exitCode=0 Oct 13 09:06:06 crc kubenswrapper[4685]: I1013 09:06:06.794856 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3b0348a6-6814-47f3-919e-e975c01e00e8","Type":"ContainerDied","Data":"692441bc08af17df5ba27504a2d82f321da33197cff11964acc63b131225301e"} Oct 13 09:06:06 crc kubenswrapper[4685]: I1013 09:06:06.795159 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3b0348a6-6814-47f3-919e-e975c01e00e8","Type":"ContainerDied","Data":"42c24f17559cb28c52a00ab2cc29b06472d9e601bebc2c8573f8bb35d9a68633"} Oct 13 09:06:06 crc kubenswrapper[4685]: I1013 09:06:06.795175 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42c24f17559cb28c52a00ab2cc29b06472d9e601bebc2c8573f8bb35d9a68633" Oct 13 09:06:06 crc kubenswrapper[4685]: I1013 09:06:06.859528 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 13 09:06:06 crc kubenswrapper[4685]: I1013 09:06:06.982967 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-plugins\") pod \"3b0348a6-6814-47f3-919e-e975c01e00e8\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " Oct 13 09:06:06 crc kubenswrapper[4685]: I1013 09:06:06.983049 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3b0348a6-6814-47f3-919e-e975c01e00e8-pod-info\") pod \"3b0348a6-6814-47f3-919e-e975c01e00e8\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " Oct 13 09:06:06 crc kubenswrapper[4685]: I1013 09:06:06.983075 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-tls\") pod \"3b0348a6-6814-47f3-919e-e975c01e00e8\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " Oct 13 09:06:06 crc kubenswrapper[4685]: I1013 09:06:06.983181 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3b0348a6-6814-47f3-919e-e975c01e00e8-erlang-cookie-secret\") pod \"3b0348a6-6814-47f3-919e-e975c01e00e8\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " Oct 13 09:06:06 crc kubenswrapper[4685]: I1013 09:06:06.983204 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3b0348a6-6814-47f3-919e-e975c01e00e8-plugins-conf\") pod \"3b0348a6-6814-47f3-919e-e975c01e00e8\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " Oct 13 09:06:06 crc kubenswrapper[4685]: I1013 09:06:06.983235 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-erlang-cookie\") pod \"3b0348a6-6814-47f3-919e-e975c01e00e8\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " Oct 13 09:06:06 crc kubenswrapper[4685]: I1013 09:06:06.983254 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-confd\") pod \"3b0348a6-6814-47f3-919e-e975c01e00e8\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " Oct 13 09:06:06 crc kubenswrapper[4685]: I1013 09:06:06.983277 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"3b0348a6-6814-47f3-919e-e975c01e00e8\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " Oct 13 09:06:06 crc kubenswrapper[4685]: I1013 09:06:06.983303 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b0348a6-6814-47f3-919e-e975c01e00e8-config-data\") pod \"3b0348a6-6814-47f3-919e-e975c01e00e8\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " Oct 13 09:06:06 crc kubenswrapper[4685]: I1013 09:06:06.983324 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3b0348a6-6814-47f3-919e-e975c01e00e8-server-conf\") pod \"3b0348a6-6814-47f3-919e-e975c01e00e8\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " Oct 13 09:06:06 crc kubenswrapper[4685]: I1013 09:06:06.983349 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4kzp\" (UniqueName: \"kubernetes.io/projected/3b0348a6-6814-47f3-919e-e975c01e00e8-kube-api-access-v4kzp\") pod \"3b0348a6-6814-47f3-919e-e975c01e00e8\" (UID: \"3b0348a6-6814-47f3-919e-e975c01e00e8\") " Oct 13 09:06:06 crc kubenswrapper[4685]: I1013 09:06:06.987727 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b0348a6-6814-47f3-919e-e975c01e00e8-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "3b0348a6-6814-47f3-919e-e975c01e00e8" (UID: "3b0348a6-6814-47f3-919e-e975c01e00e8"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:06:06 crc kubenswrapper[4685]: I1013 09:06:06.989059 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "3b0348a6-6814-47f3-919e-e975c01e00e8" (UID: "3b0348a6-6814-47f3-919e-e975c01e00e8"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:06:06 crc kubenswrapper[4685]: I1013 09:06:06.991436 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "3b0348a6-6814-47f3-919e-e975c01e00e8" (UID: "3b0348a6-6814-47f3-919e-e975c01e00e8"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:06:06 crc kubenswrapper[4685]: I1013 09:06:06.993718 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b0348a6-6814-47f3-919e-e975c01e00e8-kube-api-access-v4kzp" (OuterVolumeSpecName: "kube-api-access-v4kzp") pod "3b0348a6-6814-47f3-919e-e975c01e00e8" (UID: "3b0348a6-6814-47f3-919e-e975c01e00e8"). InnerVolumeSpecName "kube-api-access-v4kzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.005165 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/3b0348a6-6814-47f3-919e-e975c01e00e8-pod-info" (OuterVolumeSpecName: "pod-info") pod "3b0348a6-6814-47f3-919e-e975c01e00e8" (UID: "3b0348a6-6814-47f3-919e-e975c01e00e8"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.019306 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b0348a6-6814-47f3-919e-e975c01e00e8-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "3b0348a6-6814-47f3-919e-e975c01e00e8" (UID: "3b0348a6-6814-47f3-919e-e975c01e00e8"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.020165 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "persistence") pod "3b0348a6-6814-47f3-919e-e975c01e00e8" (UID: "3b0348a6-6814-47f3-919e-e975c01e00e8"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.045708 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "3b0348a6-6814-47f3-919e-e975c01e00e8" (UID: "3b0348a6-6814-47f3-919e-e975c01e00e8"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.085965 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b0348a6-6814-47f3-919e-e975c01e00e8-server-conf" (OuterVolumeSpecName: "server-conf") pod "3b0348a6-6814-47f3-919e-e975c01e00e8" (UID: "3b0348a6-6814-47f3-919e-e975c01e00e8"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.086532 4685 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3b0348a6-6814-47f3-919e-e975c01e00e8-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.086580 4685 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3b0348a6-6814-47f3-919e-e975c01e00e8-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.086594 4685 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.086622 4685 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.086633 4685 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3b0348a6-6814-47f3-919e-e975c01e00e8-server-conf\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.086644 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4kzp\" (UniqueName: \"kubernetes.io/projected/3b0348a6-6814-47f3-919e-e975c01e00e8-kube-api-access-v4kzp\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.086655 4685 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.086666 4685 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3b0348a6-6814-47f3-919e-e975c01e00e8-pod-info\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.086677 4685 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.091426 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b0348a6-6814-47f3-919e-e975c01e00e8-config-data" (OuterVolumeSpecName: "config-data") pod "3b0348a6-6814-47f3-919e-e975c01e00e8" (UID: "3b0348a6-6814-47f3-919e-e975c01e00e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.152626 4685 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.187322 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "3b0348a6-6814-47f3-919e-e975c01e00e8" (UID: "3b0348a6-6814-47f3-919e-e975c01e00e8"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.189493 4685 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3b0348a6-6814-47f3-919e-e975c01e00e8-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.189522 4685 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.189536 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b0348a6-6814-47f3-919e-e975c01e00e8-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.283724 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.290570 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-tls\") pod \"b7f19239-a2d9-49b5-80b0-43543daf1787\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.290669 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-erlang-cookie\") pod \"b7f19239-a2d9-49b5-80b0-43543daf1787\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.290695 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"b7f19239-a2d9-49b5-80b0-43543daf1787\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.290723 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b7f19239-a2d9-49b5-80b0-43543daf1787-pod-info\") pod \"b7f19239-a2d9-49b5-80b0-43543daf1787\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.290758 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b7f19239-a2d9-49b5-80b0-43543daf1787-plugins-conf\") pod \"b7f19239-a2d9-49b5-80b0-43543daf1787\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.290824 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-plugins\") pod \"b7f19239-a2d9-49b5-80b0-43543daf1787\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.290852 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-confd\") pod \"b7f19239-a2d9-49b5-80b0-43543daf1787\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.290889 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b7f19239-a2d9-49b5-80b0-43543daf1787-config-data\") pod \"b7f19239-a2d9-49b5-80b0-43543daf1787\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.290985 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sc47f\" (UniqueName: \"kubernetes.io/projected/b7f19239-a2d9-49b5-80b0-43543daf1787-kube-api-access-sc47f\") pod \"b7f19239-a2d9-49b5-80b0-43543daf1787\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.291003 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b7f19239-a2d9-49b5-80b0-43543daf1787-server-conf\") pod \"b7f19239-a2d9-49b5-80b0-43543daf1787\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.291026 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b7f19239-a2d9-49b5-80b0-43543daf1787-erlang-cookie-secret\") pod \"b7f19239-a2d9-49b5-80b0-43543daf1787\" (UID: \"b7f19239-a2d9-49b5-80b0-43543daf1787\") " Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.292661 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7f19239-a2d9-49b5-80b0-43543daf1787-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "b7f19239-a2d9-49b5-80b0-43543daf1787" (UID: "b7f19239-a2d9-49b5-80b0-43543daf1787"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.294024 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "b7f19239-a2d9-49b5-80b0-43543daf1787" (UID: "b7f19239-a2d9-49b5-80b0-43543daf1787"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.295779 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7f19239-a2d9-49b5-80b0-43543daf1787-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "b7f19239-a2d9-49b5-80b0-43543daf1787" (UID: "b7f19239-a2d9-49b5-80b0-43543daf1787"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.296973 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/b7f19239-a2d9-49b5-80b0-43543daf1787-pod-info" (OuterVolumeSpecName: "pod-info") pod "b7f19239-a2d9-49b5-80b0-43543daf1787" (UID: "b7f19239-a2d9-49b5-80b0-43543daf1787"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.300044 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "b7f19239-a2d9-49b5-80b0-43543daf1787" (UID: "b7f19239-a2d9-49b5-80b0-43543daf1787"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.301176 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7f19239-a2d9-49b5-80b0-43543daf1787-kube-api-access-sc47f" (OuterVolumeSpecName: "kube-api-access-sc47f") pod "b7f19239-a2d9-49b5-80b0-43543daf1787" (UID: "b7f19239-a2d9-49b5-80b0-43543daf1787"). InnerVolumeSpecName "kube-api-access-sc47f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.306640 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "b7f19239-a2d9-49b5-80b0-43543daf1787" (UID: "b7f19239-a2d9-49b5-80b0-43543daf1787"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.330178 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "b7f19239-a2d9-49b5-80b0-43543daf1787" (UID: "b7f19239-a2d9-49b5-80b0-43543daf1787"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.358785 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7f19239-a2d9-49b5-80b0-43543daf1787-config-data" (OuterVolumeSpecName: "config-data") pod "b7f19239-a2d9-49b5-80b0-43543daf1787" (UID: "b7f19239-a2d9-49b5-80b0-43543daf1787"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.393697 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sc47f\" (UniqueName: \"kubernetes.io/projected/b7f19239-a2d9-49b5-80b0-43543daf1787-kube-api-access-sc47f\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.393729 4685 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b7f19239-a2d9-49b5-80b0-43543daf1787-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.393738 4685 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.393751 4685 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.393780 4685 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.393789 4685 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b7f19239-a2d9-49b5-80b0-43543daf1787-pod-info\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.393797 4685 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b7f19239-a2d9-49b5-80b0-43543daf1787-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.393806 4685 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.393814 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b7f19239-a2d9-49b5-80b0-43543daf1787-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.397823 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7f19239-a2d9-49b5-80b0-43543daf1787-server-conf" (OuterVolumeSpecName: "server-conf") pod "b7f19239-a2d9-49b5-80b0-43543daf1787" (UID: "b7f19239-a2d9-49b5-80b0-43543daf1787"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.427234 4685 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.448241 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "b7f19239-a2d9-49b5-80b0-43543daf1787" (UID: "b7f19239-a2d9-49b5-80b0-43543daf1787"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.495426 4685 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b7f19239-a2d9-49b5-80b0-43543daf1787-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.495461 4685 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b7f19239-a2d9-49b5-80b0-43543daf1787-server-conf\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.495478 4685 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.805808 4685 generic.go:334] "Generic (PLEG): container finished" podID="b7f19239-a2d9-49b5-80b0-43543daf1787" containerID="0ab5c1c1eaab4696cc58a8975e88c605263b37c6525d8d978b6c52d12ae3182e" exitCode=0 Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.805859 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b7f19239-a2d9-49b5-80b0-43543daf1787","Type":"ContainerDied","Data":"0ab5c1c1eaab4696cc58a8975e88c605263b37c6525d8d978b6c52d12ae3182e"} Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.806212 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.806230 4685 scope.go:117] "RemoveContainer" containerID="0ab5c1c1eaab4696cc58a8975e88c605263b37c6525d8d978b6c52d12ae3182e" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.805893 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.806207 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b7f19239-a2d9-49b5-80b0-43543daf1787","Type":"ContainerDied","Data":"4355a58fb21aefb60a8f8b4cff96124de1105ec1603c270f7504162362dcbf32"} Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.830968 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.832620 4685 scope.go:117] "RemoveContainer" containerID="a6d714e0fc56e7256e93150fb3ec702f4835022733af8f3f1d12eecd1ec2d20b" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.840978 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.866981 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.874957 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.891792 4685 scope.go:117] "RemoveContainer" containerID="0ab5c1c1eaab4696cc58a8975e88c605263b37c6525d8d978b6c52d12ae3182e" Oct 13 09:06:07 crc kubenswrapper[4685]: E1013 09:06:07.892376 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ab5c1c1eaab4696cc58a8975e88c605263b37c6525d8d978b6c52d12ae3182e\": container with ID starting with 0ab5c1c1eaab4696cc58a8975e88c605263b37c6525d8d978b6c52d12ae3182e not found: ID does not exist" containerID="0ab5c1c1eaab4696cc58a8975e88c605263b37c6525d8d978b6c52d12ae3182e" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.892429 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ab5c1c1eaab4696cc58a8975e88c605263b37c6525d8d978b6c52d12ae3182e"} err="failed to get container status \"0ab5c1c1eaab4696cc58a8975e88c605263b37c6525d8d978b6c52d12ae3182e\": rpc error: code = NotFound desc = could not find container \"0ab5c1c1eaab4696cc58a8975e88c605263b37c6525d8d978b6c52d12ae3182e\": container with ID starting with 0ab5c1c1eaab4696cc58a8975e88c605263b37c6525d8d978b6c52d12ae3182e not found: ID does not exist" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.892461 4685 scope.go:117] "RemoveContainer" containerID="a6d714e0fc56e7256e93150fb3ec702f4835022733af8f3f1d12eecd1ec2d20b" Oct 13 09:06:07 crc kubenswrapper[4685]: E1013 09:06:07.903420 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6d714e0fc56e7256e93150fb3ec702f4835022733af8f3f1d12eecd1ec2d20b\": container with ID starting with a6d714e0fc56e7256e93150fb3ec702f4835022733af8f3f1d12eecd1ec2d20b not found: ID does not exist" containerID="a6d714e0fc56e7256e93150fb3ec702f4835022733af8f3f1d12eecd1ec2d20b" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.903490 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6d714e0fc56e7256e93150fb3ec702f4835022733af8f3f1d12eecd1ec2d20b"} err="failed to get container status \"a6d714e0fc56e7256e93150fb3ec702f4835022733af8f3f1d12eecd1ec2d20b\": rpc error: code = NotFound desc = could not find container \"a6d714e0fc56e7256e93150fb3ec702f4835022733af8f3f1d12eecd1ec2d20b\": container with ID starting with a6d714e0fc56e7256e93150fb3ec702f4835022733af8f3f1d12eecd1ec2d20b not found: ID does not exist" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.909092 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 13 09:06:07 crc kubenswrapper[4685]: E1013 09:06:07.910423 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7f19239-a2d9-49b5-80b0-43543daf1787" containerName="rabbitmq" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.911100 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7f19239-a2d9-49b5-80b0-43543daf1787" containerName="rabbitmq" Oct 13 09:06:07 crc kubenswrapper[4685]: E1013 09:06:07.911196 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b0348a6-6814-47f3-919e-e975c01e00e8" containerName="setup-container" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.911246 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b0348a6-6814-47f3-919e-e975c01e00e8" containerName="setup-container" Oct 13 09:06:07 crc kubenswrapper[4685]: E1013 09:06:07.911324 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7f19239-a2d9-49b5-80b0-43543daf1787" containerName="setup-container" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.911372 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7f19239-a2d9-49b5-80b0-43543daf1787" containerName="setup-container" Oct 13 09:06:07 crc kubenswrapper[4685]: E1013 09:06:07.911432 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b0348a6-6814-47f3-919e-e975c01e00e8" containerName="rabbitmq" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.911488 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b0348a6-6814-47f3-919e-e975c01e00e8" containerName="rabbitmq" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.911942 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b0348a6-6814-47f3-919e-e975c01e00e8" containerName="rabbitmq" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.914438 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7f19239-a2d9-49b5-80b0-43543daf1787" containerName="rabbitmq" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.918134 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.942150 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.942357 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-5c4fv" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.943524 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.943669 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.944011 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.952235 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.952493 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.969901 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.971640 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.974599 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.974786 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.974955 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-b9zgt" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.975099 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.975854 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.976128 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 13 09:06:07 crc kubenswrapper[4685]: I1013 09:06:07.976310 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.015017 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.018998 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.121948 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.122157 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/27720755-e830-4eb4-b0e0-b5dfe9ceb253-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.122265 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.122370 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.122442 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/27720755-e830-4eb4-b0e0-b5dfe9ceb253-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.122508 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.122572 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.122645 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/27720755-e830-4eb4-b0e0-b5dfe9ceb253-pod-info\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.122736 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/27720755-e830-4eb4-b0e0-b5dfe9ceb253-server-conf\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.122763 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/27720755-e830-4eb4-b0e0-b5dfe9ceb253-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.122796 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7lmx\" (UniqueName: \"kubernetes.io/projected/27720755-e830-4eb4-b0e0-b5dfe9ceb253-kube-api-access-m7lmx\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.122820 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/27720755-e830-4eb4-b0e0-b5dfe9ceb253-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.122864 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.122883 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.122935 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/27720755-e830-4eb4-b0e0-b5dfe9ceb253-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.122962 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.122982 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.122999 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/27720755-e830-4eb4-b0e0-b5dfe9ceb253-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.123157 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t6sh\" (UniqueName: \"kubernetes.io/projected/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-kube-api-access-9t6sh\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.123222 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27720755-e830-4eb4-b0e0-b5dfe9ceb253-config-data\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.123242 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.123286 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.225458 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t6sh\" (UniqueName: \"kubernetes.io/projected/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-kube-api-access-9t6sh\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.225525 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27720755-e830-4eb4-b0e0-b5dfe9ceb253-config-data\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.225545 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.226432 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27720755-e830-4eb4-b0e0-b5dfe9ceb253-config-data\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.226494 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.227173 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.227230 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.227469 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.228879 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/27720755-e830-4eb4-b0e0-b5dfe9ceb253-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.228906 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.228948 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.228965 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/27720755-e830-4eb4-b0e0-b5dfe9ceb253-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.228986 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.229001 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/27720755-e830-4eb4-b0e0-b5dfe9ceb253-pod-info\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.229017 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.229038 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/27720755-e830-4eb4-b0e0-b5dfe9ceb253-server-conf\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.229057 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/27720755-e830-4eb4-b0e0-b5dfe9ceb253-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.230094 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.232789 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.233157 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.234997 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.235015 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/27720755-e830-4eb4-b0e0-b5dfe9ceb253-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.235622 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/27720755-e830-4eb4-b0e0-b5dfe9ceb253-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.235718 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/27720755-e830-4eb4-b0e0-b5dfe9ceb253-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.236041 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7lmx\" (UniqueName: \"kubernetes.io/projected/27720755-e830-4eb4-b0e0-b5dfe9ceb253-kube-api-access-m7lmx\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.236414 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/27720755-e830-4eb4-b0e0-b5dfe9ceb253-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.236490 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.236523 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.236595 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/27720755-e830-4eb4-b0e0-b5dfe9ceb253-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.236646 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.236671 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.236687 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/27720755-e830-4eb4-b0e0-b5dfe9ceb253-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.237693 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/27720755-e830-4eb4-b0e0-b5dfe9ceb253-pod-info\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.237850 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/27720755-e830-4eb4-b0e0-b5dfe9ceb253-server-conf\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.238812 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.239208 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.239316 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.240546 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/27720755-e830-4eb4-b0e0-b5dfe9ceb253-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.242220 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/27720755-e830-4eb4-b0e0-b5dfe9ceb253-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.254994 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.257820 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.258355 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/27720755-e830-4eb4-b0e0-b5dfe9ceb253-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.262149 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t6sh\" (UniqueName: \"kubernetes.io/projected/5f68bc3a-ebea-44a6-9b00-048e6afd1d09-kube-api-access-9t6sh\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.268811 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7lmx\" (UniqueName: \"kubernetes.io/projected/27720755-e830-4eb4-b0e0-b5dfe9ceb253-kube-api-access-m7lmx\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.287742 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"27720755-e830-4eb4-b0e0-b5dfe9ceb253\") " pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.291844 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f68bc3a-ebea-44a6-9b00-048e6afd1d09\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.305398 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.573603 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.820060 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"27720755-e830-4eb4-b0e0-b5dfe9ceb253","Type":"ContainerStarted","Data":"dd628a76abad2fba396c0cb796c39bf9b4cc87cd45e5a516fe8af235ff3513e7"} Oct 13 09:06:08 crc kubenswrapper[4685]: I1013 09:06:08.833379 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 13 09:06:09 crc kubenswrapper[4685]: I1013 09:06:09.035309 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 13 09:06:09 crc kubenswrapper[4685]: I1013 09:06:09.520678 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b0348a6-6814-47f3-919e-e975c01e00e8" path="/var/lib/kubelet/pods/3b0348a6-6814-47f3-919e-e975c01e00e8/volumes" Oct 13 09:06:09 crc kubenswrapper[4685]: I1013 09:06:09.521455 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7f19239-a2d9-49b5-80b0-43543daf1787" path="/var/lib/kubelet/pods/b7f19239-a2d9-49b5-80b0-43543daf1787/volumes" Oct 13 09:06:09 crc kubenswrapper[4685]: I1013 09:06:09.829537 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5f68bc3a-ebea-44a6-9b00-048e6afd1d09","Type":"ContainerStarted","Data":"b507297baa29c31b9de01e0468fd82e91d93cff6c010efff2353f1dfbc8b1508"} Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.426075 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d558885bc-l772x"] Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.430093 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.435114 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.458932 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-l772x"] Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.579646 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-config\") pod \"dnsmasq-dns-d558885bc-l772x\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.579909 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-l772x\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.580008 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-l772x\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.580101 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-dns-svc\") pod \"dnsmasq-dns-d558885bc-l772x\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.580203 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ssln\" (UniqueName: \"kubernetes.io/projected/1b46cbc3-a4eb-4e31-8939-139e60849422-kube-api-access-7ssln\") pod \"dnsmasq-dns-d558885bc-l772x\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.580389 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-l772x\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.580476 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-l772x\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.682687 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-l772x\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.682744 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-l772x\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.682796 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-dns-svc\") pod \"dnsmasq-dns-d558885bc-l772x\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.682827 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ssln\" (UniqueName: \"kubernetes.io/projected/1b46cbc3-a4eb-4e31-8939-139e60849422-kube-api-access-7ssln\") pod \"dnsmasq-dns-d558885bc-l772x\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.682994 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-l772x\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.683042 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-l772x\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.683068 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-config\") pod \"dnsmasq-dns-d558885bc-l772x\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.683982 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-config\") pod \"dnsmasq-dns-d558885bc-l772x\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.684234 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-l772x\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.684356 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-l772x\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.684389 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-dns-svc\") pod \"dnsmasq-dns-d558885bc-l772x\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.684457 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-l772x\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.685448 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-l772x\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.707889 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ssln\" (UniqueName: \"kubernetes.io/projected/1b46cbc3-a4eb-4e31-8939-139e60849422-kube-api-access-7ssln\") pod \"dnsmasq-dns-d558885bc-l772x\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.765569 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.854512 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"27720755-e830-4eb4-b0e0-b5dfe9ceb253","Type":"ContainerStarted","Data":"01e3070b5eb17ee80b432adefb1f1b746fade1084eda0cbdde5295d30cc6c635"} Oct 13 09:06:10 crc kubenswrapper[4685]: I1013 09:06:10.861522 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5f68bc3a-ebea-44a6-9b00-048e6afd1d09","Type":"ContainerStarted","Data":"8852ef5d9272e4c6e57d9fd67acfe5a8ca850ae6207f58ca4f082ec8afdf85cb"} Oct 13 09:06:11 crc kubenswrapper[4685]: I1013 09:06:11.268394 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-l772x"] Oct 13 09:06:11 crc kubenswrapper[4685]: I1013 09:06:11.874856 4685 generic.go:334] "Generic (PLEG): container finished" podID="1b46cbc3-a4eb-4e31-8939-139e60849422" containerID="f861bdd5febdca438122196cf83dbef1443fe6b03d5e9eaddbd522801258b61b" exitCode=0 Oct 13 09:06:11 crc kubenswrapper[4685]: I1013 09:06:11.875107 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-l772x" event={"ID":"1b46cbc3-a4eb-4e31-8939-139e60849422","Type":"ContainerDied","Data":"f861bdd5febdca438122196cf83dbef1443fe6b03d5e9eaddbd522801258b61b"} Oct 13 09:06:11 crc kubenswrapper[4685]: I1013 09:06:11.875357 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-l772x" event={"ID":"1b46cbc3-a4eb-4e31-8939-139e60849422","Type":"ContainerStarted","Data":"30a918ecf2e7c15f2854b45ce7782a85d531d979c1322d4c0831cc21df10babb"} Oct 13 09:06:12 crc kubenswrapper[4685]: I1013 09:06:12.888025 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-l772x" event={"ID":"1b46cbc3-a4eb-4e31-8939-139e60849422","Type":"ContainerStarted","Data":"a62145b7126d2fee2a82c50032c4675754fcada695f665b30023b65b76a68b69"} Oct 13 09:06:12 crc kubenswrapper[4685]: I1013 09:06:12.888521 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:12 crc kubenswrapper[4685]: I1013 09:06:12.912192 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d558885bc-l772x" podStartSLOduration=2.912173849 podStartE2EDuration="2.912173849s" podCreationTimestamp="2025-10-13 09:06:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:06:12.908044017 +0000 UTC m=+1298.055919778" watchObservedRunningTime="2025-10-13 09:06:12.912173849 +0000 UTC m=+1298.060049610" Oct 13 09:06:20 crc kubenswrapper[4685]: I1013 09:06:20.766992 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:20 crc kubenswrapper[4685]: I1013 09:06:20.835598 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-dnzbg"] Oct 13 09:06:20 crc kubenswrapper[4685]: I1013 09:06:20.835880 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" podUID="dfd5d590-0f0f-4de4-a373-ac9311796783" containerName="dnsmasq-dns" containerID="cri-o://82f9ced2205498c22984feb3b2c5ac6396a247b6b34b47113c6872c93a04973d" gracePeriod=10 Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.029460 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b865b64bc-nsk84"] Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.040504 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.050085 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b865b64bc-nsk84"] Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.095262 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/908f2660-8000-46f8-8119-e43b00f79abb-dns-svc\") pod \"dnsmasq-dns-6b865b64bc-nsk84\" (UID: \"908f2660-8000-46f8-8119-e43b00f79abb\") " pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.095320 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l87q\" (UniqueName: \"kubernetes.io/projected/908f2660-8000-46f8-8119-e43b00f79abb-kube-api-access-8l87q\") pod \"dnsmasq-dns-6b865b64bc-nsk84\" (UID: \"908f2660-8000-46f8-8119-e43b00f79abb\") " pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.095373 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/908f2660-8000-46f8-8119-e43b00f79abb-config\") pod \"dnsmasq-dns-6b865b64bc-nsk84\" (UID: \"908f2660-8000-46f8-8119-e43b00f79abb\") " pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.095397 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/908f2660-8000-46f8-8119-e43b00f79abb-openstack-edpm-ipam\") pod \"dnsmasq-dns-6b865b64bc-nsk84\" (UID: \"908f2660-8000-46f8-8119-e43b00f79abb\") " pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.095418 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/908f2660-8000-46f8-8119-e43b00f79abb-ovsdbserver-nb\") pod \"dnsmasq-dns-6b865b64bc-nsk84\" (UID: \"908f2660-8000-46f8-8119-e43b00f79abb\") " pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.095484 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/908f2660-8000-46f8-8119-e43b00f79abb-dns-swift-storage-0\") pod \"dnsmasq-dns-6b865b64bc-nsk84\" (UID: \"908f2660-8000-46f8-8119-e43b00f79abb\") " pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.095649 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/908f2660-8000-46f8-8119-e43b00f79abb-ovsdbserver-sb\") pod \"dnsmasq-dns-6b865b64bc-nsk84\" (UID: \"908f2660-8000-46f8-8119-e43b00f79abb\") " pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.197304 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/908f2660-8000-46f8-8119-e43b00f79abb-dns-swift-storage-0\") pod \"dnsmasq-dns-6b865b64bc-nsk84\" (UID: \"908f2660-8000-46f8-8119-e43b00f79abb\") " pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.197391 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/908f2660-8000-46f8-8119-e43b00f79abb-ovsdbserver-sb\") pod \"dnsmasq-dns-6b865b64bc-nsk84\" (UID: \"908f2660-8000-46f8-8119-e43b00f79abb\") " pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.197447 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/908f2660-8000-46f8-8119-e43b00f79abb-dns-svc\") pod \"dnsmasq-dns-6b865b64bc-nsk84\" (UID: \"908f2660-8000-46f8-8119-e43b00f79abb\") " pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.197478 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l87q\" (UniqueName: \"kubernetes.io/projected/908f2660-8000-46f8-8119-e43b00f79abb-kube-api-access-8l87q\") pod \"dnsmasq-dns-6b865b64bc-nsk84\" (UID: \"908f2660-8000-46f8-8119-e43b00f79abb\") " pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.197526 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/908f2660-8000-46f8-8119-e43b00f79abb-config\") pod \"dnsmasq-dns-6b865b64bc-nsk84\" (UID: \"908f2660-8000-46f8-8119-e43b00f79abb\") " pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.197549 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/908f2660-8000-46f8-8119-e43b00f79abb-openstack-edpm-ipam\") pod \"dnsmasq-dns-6b865b64bc-nsk84\" (UID: \"908f2660-8000-46f8-8119-e43b00f79abb\") " pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.197569 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/908f2660-8000-46f8-8119-e43b00f79abb-ovsdbserver-nb\") pod \"dnsmasq-dns-6b865b64bc-nsk84\" (UID: \"908f2660-8000-46f8-8119-e43b00f79abb\") " pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.198420 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/908f2660-8000-46f8-8119-e43b00f79abb-ovsdbserver-nb\") pod \"dnsmasq-dns-6b865b64bc-nsk84\" (UID: \"908f2660-8000-46f8-8119-e43b00f79abb\") " pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.199000 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/908f2660-8000-46f8-8119-e43b00f79abb-dns-swift-storage-0\") pod \"dnsmasq-dns-6b865b64bc-nsk84\" (UID: \"908f2660-8000-46f8-8119-e43b00f79abb\") " pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.199482 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/908f2660-8000-46f8-8119-e43b00f79abb-ovsdbserver-sb\") pod \"dnsmasq-dns-6b865b64bc-nsk84\" (UID: \"908f2660-8000-46f8-8119-e43b00f79abb\") " pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.200093 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/908f2660-8000-46f8-8119-e43b00f79abb-dns-svc\") pod \"dnsmasq-dns-6b865b64bc-nsk84\" (UID: \"908f2660-8000-46f8-8119-e43b00f79abb\") " pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.200836 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/908f2660-8000-46f8-8119-e43b00f79abb-config\") pod \"dnsmasq-dns-6b865b64bc-nsk84\" (UID: \"908f2660-8000-46f8-8119-e43b00f79abb\") " pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.201359 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/908f2660-8000-46f8-8119-e43b00f79abb-openstack-edpm-ipam\") pod \"dnsmasq-dns-6b865b64bc-nsk84\" (UID: \"908f2660-8000-46f8-8119-e43b00f79abb\") " pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.241212 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l87q\" (UniqueName: \"kubernetes.io/projected/908f2660-8000-46f8-8119-e43b00f79abb-kube-api-access-8l87q\") pod \"dnsmasq-dns-6b865b64bc-nsk84\" (UID: \"908f2660-8000-46f8-8119-e43b00f79abb\") " pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.413020 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.848068 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.850866 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b865b64bc-nsk84"] Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.910584 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pznk8\" (UniqueName: \"kubernetes.io/projected/dfd5d590-0f0f-4de4-a373-ac9311796783-kube-api-access-pznk8\") pod \"dfd5d590-0f0f-4de4-a373-ac9311796783\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.910775 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-ovsdbserver-sb\") pod \"dfd5d590-0f0f-4de4-a373-ac9311796783\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.910812 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-ovsdbserver-nb\") pod \"dfd5d590-0f0f-4de4-a373-ac9311796783\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.910868 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-dns-svc\") pod \"dfd5d590-0f0f-4de4-a373-ac9311796783\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.910899 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-config\") pod \"dfd5d590-0f0f-4de4-a373-ac9311796783\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.910971 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-dns-swift-storage-0\") pod \"dfd5d590-0f0f-4de4-a373-ac9311796783\" (UID: \"dfd5d590-0f0f-4de4-a373-ac9311796783\") " Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.932182 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfd5d590-0f0f-4de4-a373-ac9311796783-kube-api-access-pznk8" (OuterVolumeSpecName: "kube-api-access-pznk8") pod "dfd5d590-0f0f-4de4-a373-ac9311796783" (UID: "dfd5d590-0f0f-4de4-a373-ac9311796783"). InnerVolumeSpecName "kube-api-access-pznk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.972617 4685 generic.go:334] "Generic (PLEG): container finished" podID="dfd5d590-0f0f-4de4-a373-ac9311796783" containerID="82f9ced2205498c22984feb3b2c5ac6396a247b6b34b47113c6872c93a04973d" exitCode=0 Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.972826 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" event={"ID":"dfd5d590-0f0f-4de4-a373-ac9311796783","Type":"ContainerDied","Data":"82f9ced2205498c22984feb3b2c5ac6396a247b6b34b47113c6872c93a04973d"} Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.972903 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" event={"ID":"dfd5d590-0f0f-4de4-a373-ac9311796783","Type":"ContainerDied","Data":"698641b4ca252630d5694da66cb87f513b3864f1f768e5acfa57087956b06b59"} Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.973168 4685 scope.go:117] "RemoveContainer" containerID="82f9ced2205498c22984feb3b2c5ac6396a247b6b34b47113c6872c93a04973d" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.973342 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-dnzbg" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.980458 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" event={"ID":"908f2660-8000-46f8-8119-e43b00f79abb","Type":"ContainerStarted","Data":"8ddcac102074b26d324b9a9148f87ee6c8fc5d82a486cd6f64f27ddbf2d14874"} Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.990390 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "dfd5d590-0f0f-4de4-a373-ac9311796783" (UID: "dfd5d590-0f0f-4de4-a373-ac9311796783"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.993972 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "dfd5d590-0f0f-4de4-a373-ac9311796783" (UID: "dfd5d590-0f0f-4de4-a373-ac9311796783"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:06:21 crc kubenswrapper[4685]: I1013 09:06:21.997067 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dfd5d590-0f0f-4de4-a373-ac9311796783" (UID: "dfd5d590-0f0f-4de4-a373-ac9311796783"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:06:22 crc kubenswrapper[4685]: I1013 09:06:22.003616 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-config" (OuterVolumeSpecName: "config") pod "dfd5d590-0f0f-4de4-a373-ac9311796783" (UID: "dfd5d590-0f0f-4de4-a373-ac9311796783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:06:22 crc kubenswrapper[4685]: I1013 09:06:22.013011 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dfd5d590-0f0f-4de4-a373-ac9311796783" (UID: "dfd5d590-0f0f-4de4-a373-ac9311796783"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:06:22 crc kubenswrapper[4685]: I1013 09:06:22.013469 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:22 crc kubenswrapper[4685]: I1013 09:06:22.013554 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:22 crc kubenswrapper[4685]: I1013 09:06:22.013622 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:22 crc kubenswrapper[4685]: I1013 09:06:22.013689 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-config\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:22 crc kubenswrapper[4685]: I1013 09:06:22.013754 4685 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dfd5d590-0f0f-4de4-a373-ac9311796783-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:22 crc kubenswrapper[4685]: I1013 09:06:22.013819 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pznk8\" (UniqueName: \"kubernetes.io/projected/dfd5d590-0f0f-4de4-a373-ac9311796783-kube-api-access-pznk8\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:22 crc kubenswrapper[4685]: I1013 09:06:22.102820 4685 scope.go:117] "RemoveContainer" containerID="e05dc373d653e3abdc9fe349a4a33bafda1bb2555a91c854b35b3c4b4325a766" Oct 13 09:06:22 crc kubenswrapper[4685]: I1013 09:06:22.127727 4685 scope.go:117] "RemoveContainer" containerID="82f9ced2205498c22984feb3b2c5ac6396a247b6b34b47113c6872c93a04973d" Oct 13 09:06:22 crc kubenswrapper[4685]: E1013 09:06:22.130196 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82f9ced2205498c22984feb3b2c5ac6396a247b6b34b47113c6872c93a04973d\": container with ID starting with 82f9ced2205498c22984feb3b2c5ac6396a247b6b34b47113c6872c93a04973d not found: ID does not exist" containerID="82f9ced2205498c22984feb3b2c5ac6396a247b6b34b47113c6872c93a04973d" Oct 13 09:06:22 crc kubenswrapper[4685]: I1013 09:06:22.130248 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82f9ced2205498c22984feb3b2c5ac6396a247b6b34b47113c6872c93a04973d"} err="failed to get container status \"82f9ced2205498c22984feb3b2c5ac6396a247b6b34b47113c6872c93a04973d\": rpc error: code = NotFound desc = could not find container \"82f9ced2205498c22984feb3b2c5ac6396a247b6b34b47113c6872c93a04973d\": container with ID starting with 82f9ced2205498c22984feb3b2c5ac6396a247b6b34b47113c6872c93a04973d not found: ID does not exist" Oct 13 09:06:22 crc kubenswrapper[4685]: I1013 09:06:22.130282 4685 scope.go:117] "RemoveContainer" containerID="e05dc373d653e3abdc9fe349a4a33bafda1bb2555a91c854b35b3c4b4325a766" Oct 13 09:06:22 crc kubenswrapper[4685]: E1013 09:06:22.130701 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e05dc373d653e3abdc9fe349a4a33bafda1bb2555a91c854b35b3c4b4325a766\": container with ID starting with e05dc373d653e3abdc9fe349a4a33bafda1bb2555a91c854b35b3c4b4325a766 not found: ID does not exist" containerID="e05dc373d653e3abdc9fe349a4a33bafda1bb2555a91c854b35b3c4b4325a766" Oct 13 09:06:22 crc kubenswrapper[4685]: I1013 09:06:22.130749 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e05dc373d653e3abdc9fe349a4a33bafda1bb2555a91c854b35b3c4b4325a766"} err="failed to get container status \"e05dc373d653e3abdc9fe349a4a33bafda1bb2555a91c854b35b3c4b4325a766\": rpc error: code = NotFound desc = could not find container \"e05dc373d653e3abdc9fe349a4a33bafda1bb2555a91c854b35b3c4b4325a766\": container with ID starting with e05dc373d653e3abdc9fe349a4a33bafda1bb2555a91c854b35b3c4b4325a766 not found: ID does not exist" Oct 13 09:06:22 crc kubenswrapper[4685]: I1013 09:06:22.310205 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-dnzbg"] Oct 13 09:06:22 crc kubenswrapper[4685]: I1013 09:06:22.318867 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-dnzbg"] Oct 13 09:06:22 crc kubenswrapper[4685]: I1013 09:06:22.992189 4685 generic.go:334] "Generic (PLEG): container finished" podID="908f2660-8000-46f8-8119-e43b00f79abb" containerID="88e3c5a2d5916a513d697bac8f57948eabe37df9a99571cf0b4a9fe2aa16712e" exitCode=0 Oct 13 09:06:22 crc kubenswrapper[4685]: I1013 09:06:22.992246 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" event={"ID":"908f2660-8000-46f8-8119-e43b00f79abb","Type":"ContainerDied","Data":"88e3c5a2d5916a513d697bac8f57948eabe37df9a99571cf0b4a9fe2aa16712e"} Oct 13 09:06:23 crc kubenswrapper[4685]: I1013 09:06:23.518038 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfd5d590-0f0f-4de4-a373-ac9311796783" path="/var/lib/kubelet/pods/dfd5d590-0f0f-4de4-a373-ac9311796783/volumes" Oct 13 09:06:24 crc kubenswrapper[4685]: I1013 09:06:24.004929 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" event={"ID":"908f2660-8000-46f8-8119-e43b00f79abb","Type":"ContainerStarted","Data":"82b8d831ca12a1db60582e2a907600de14e435b0c0141f371f882c094b31484d"} Oct 13 09:06:24 crc kubenswrapper[4685]: I1013 09:06:24.005437 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:31 crc kubenswrapper[4685]: I1013 09:06:31.414295 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" Oct 13 09:06:31 crc kubenswrapper[4685]: I1013 09:06:31.438058 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b865b64bc-nsk84" podStartSLOduration=10.438036513 podStartE2EDuration="10.438036513s" podCreationTimestamp="2025-10-13 09:06:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:06:24.032390916 +0000 UTC m=+1309.180266707" watchObservedRunningTime="2025-10-13 09:06:31.438036513 +0000 UTC m=+1316.585912274" Oct 13 09:06:31 crc kubenswrapper[4685]: I1013 09:06:31.526725 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-l772x"] Oct 13 09:06:31 crc kubenswrapper[4685]: I1013 09:06:31.526953 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d558885bc-l772x" podUID="1b46cbc3-a4eb-4e31-8939-139e60849422" containerName="dnsmasq-dns" containerID="cri-o://a62145b7126d2fee2a82c50032c4675754fcada695f665b30023b65b76a68b69" gracePeriod=10 Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.080576 4685 generic.go:334] "Generic (PLEG): container finished" podID="1b46cbc3-a4eb-4e31-8939-139e60849422" containerID="a62145b7126d2fee2a82c50032c4675754fcada695f665b30023b65b76a68b69" exitCode=0 Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.080666 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-l772x" event={"ID":"1b46cbc3-a4eb-4e31-8939-139e60849422","Type":"ContainerDied","Data":"a62145b7126d2fee2a82c50032c4675754fcada695f665b30023b65b76a68b69"} Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.467560 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.564506 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-ovsdbserver-nb\") pod \"1b46cbc3-a4eb-4e31-8939-139e60849422\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.564602 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-openstack-edpm-ipam\") pod \"1b46cbc3-a4eb-4e31-8939-139e60849422\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.564685 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-config\") pod \"1b46cbc3-a4eb-4e31-8939-139e60849422\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.564727 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ssln\" (UniqueName: \"kubernetes.io/projected/1b46cbc3-a4eb-4e31-8939-139e60849422-kube-api-access-7ssln\") pod \"1b46cbc3-a4eb-4e31-8939-139e60849422\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.564768 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-dns-svc\") pod \"1b46cbc3-a4eb-4e31-8939-139e60849422\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.564808 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-ovsdbserver-sb\") pod \"1b46cbc3-a4eb-4e31-8939-139e60849422\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.564853 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-dns-swift-storage-0\") pod \"1b46cbc3-a4eb-4e31-8939-139e60849422\" (UID: \"1b46cbc3-a4eb-4e31-8939-139e60849422\") " Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.585254 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b46cbc3-a4eb-4e31-8939-139e60849422-kube-api-access-7ssln" (OuterVolumeSpecName: "kube-api-access-7ssln") pod "1b46cbc3-a4eb-4e31-8939-139e60849422" (UID: "1b46cbc3-a4eb-4e31-8939-139e60849422"). InnerVolumeSpecName "kube-api-access-7ssln". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.668112 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ssln\" (UniqueName: \"kubernetes.io/projected/1b46cbc3-a4eb-4e31-8939-139e60849422-kube-api-access-7ssln\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.674270 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1b46cbc3-a4eb-4e31-8939-139e60849422" (UID: "1b46cbc3-a4eb-4e31-8939-139e60849422"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.678106 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1b46cbc3-a4eb-4e31-8939-139e60849422" (UID: "1b46cbc3-a4eb-4e31-8939-139e60849422"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.680959 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-config" (OuterVolumeSpecName: "config") pod "1b46cbc3-a4eb-4e31-8939-139e60849422" (UID: "1b46cbc3-a4eb-4e31-8939-139e60849422"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.686827 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1b46cbc3-a4eb-4e31-8939-139e60849422" (UID: "1b46cbc3-a4eb-4e31-8939-139e60849422"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.698450 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1b46cbc3-a4eb-4e31-8939-139e60849422" (UID: "1b46cbc3-a4eb-4e31-8939-139e60849422"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.706505 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "1b46cbc3-a4eb-4e31-8939-139e60849422" (UID: "1b46cbc3-a4eb-4e31-8939-139e60849422"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.769555 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.769590 4685 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.769600 4685 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-config\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.769609 4685 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.769617 4685 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:32 crc kubenswrapper[4685]: I1013 09:06:32.769625 4685 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b46cbc3-a4eb-4e31-8939-139e60849422-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 13 09:06:33 crc kubenswrapper[4685]: I1013 09:06:33.091768 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-l772x" Oct 13 09:06:33 crc kubenswrapper[4685]: I1013 09:06:33.091675 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-l772x" event={"ID":"1b46cbc3-a4eb-4e31-8939-139e60849422","Type":"ContainerDied","Data":"30a918ecf2e7c15f2854b45ce7782a85d531d979c1322d4c0831cc21df10babb"} Oct 13 09:06:33 crc kubenswrapper[4685]: I1013 09:06:33.096112 4685 scope.go:117] "RemoveContainer" containerID="a62145b7126d2fee2a82c50032c4675754fcada695f665b30023b65b76a68b69" Oct 13 09:06:33 crc kubenswrapper[4685]: I1013 09:06:33.122986 4685 scope.go:117] "RemoveContainer" containerID="f861bdd5febdca438122196cf83dbef1443fe6b03d5e9eaddbd522801258b61b" Oct 13 09:06:33 crc kubenswrapper[4685]: I1013 09:06:33.130422 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-l772x"] Oct 13 09:06:33 crc kubenswrapper[4685]: I1013 09:06:33.156519 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-l772x"] Oct 13 09:06:33 crc kubenswrapper[4685]: I1013 09:06:33.551966 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b46cbc3-a4eb-4e31-8939-139e60849422" path="/var/lib/kubelet/pods/1b46cbc3-a4eb-4e31-8939-139e60849422/volumes" Oct 13 09:06:43 crc kubenswrapper[4685]: I1013 09:06:43.195181 4685 generic.go:334] "Generic (PLEG): container finished" podID="5f68bc3a-ebea-44a6-9b00-048e6afd1d09" containerID="8852ef5d9272e4c6e57d9fd67acfe5a8ca850ae6207f58ca4f082ec8afdf85cb" exitCode=0 Oct 13 09:06:43 crc kubenswrapper[4685]: I1013 09:06:43.195265 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5f68bc3a-ebea-44a6-9b00-048e6afd1d09","Type":"ContainerDied","Data":"8852ef5d9272e4c6e57d9fd67acfe5a8ca850ae6207f58ca4f082ec8afdf85cb"} Oct 13 09:06:43 crc kubenswrapper[4685]: I1013 09:06:43.197667 4685 generic.go:334] "Generic (PLEG): container finished" podID="27720755-e830-4eb4-b0e0-b5dfe9ceb253" containerID="01e3070b5eb17ee80b432adefb1f1b746fade1084eda0cbdde5295d30cc6c635" exitCode=0 Oct 13 09:06:43 crc kubenswrapper[4685]: I1013 09:06:43.197702 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"27720755-e830-4eb4-b0e0-b5dfe9ceb253","Type":"ContainerDied","Data":"01e3070b5eb17ee80b432adefb1f1b746fade1084eda0cbdde5295d30cc6c635"} Oct 13 09:06:44 crc kubenswrapper[4685]: I1013 09:06:44.211039 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5f68bc3a-ebea-44a6-9b00-048e6afd1d09","Type":"ContainerStarted","Data":"b82e114e5fadfed0b4e21702bef7c317384295b3a7667fc5a5dffb953827be49"} Oct 13 09:06:44 crc kubenswrapper[4685]: I1013 09:06:44.211571 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:06:44 crc kubenswrapper[4685]: I1013 09:06:44.214056 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"27720755-e830-4eb4-b0e0-b5dfe9ceb253","Type":"ContainerStarted","Data":"fecdaa38e281e3be4c9261f692bbf449c200376f53d7b758e232d35f387c3afd"} Oct 13 09:06:44 crc kubenswrapper[4685]: I1013 09:06:44.214612 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 13 09:06:44 crc kubenswrapper[4685]: I1013 09:06:44.245015 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.244994643 podStartE2EDuration="37.244994643s" podCreationTimestamp="2025-10-13 09:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:06:44.238455295 +0000 UTC m=+1329.386331086" watchObservedRunningTime="2025-10-13 09:06:44.244994643 +0000 UTC m=+1329.392870404" Oct 13 09:06:44 crc kubenswrapper[4685]: I1013 09:06:44.268040 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.268018038 podStartE2EDuration="37.268018038s" podCreationTimestamp="2025-10-13 09:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:06:44.258662964 +0000 UTC m=+1329.406538725" watchObservedRunningTime="2025-10-13 09:06:44.268018038 +0000 UTC m=+1329.415893809" Oct 13 09:06:49 crc kubenswrapper[4685]: I1013 09:06:49.729612 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc"] Oct 13 09:06:49 crc kubenswrapper[4685]: E1013 09:06:49.730632 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfd5d590-0f0f-4de4-a373-ac9311796783" containerName="dnsmasq-dns" Oct 13 09:06:49 crc kubenswrapper[4685]: I1013 09:06:49.730649 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfd5d590-0f0f-4de4-a373-ac9311796783" containerName="dnsmasq-dns" Oct 13 09:06:49 crc kubenswrapper[4685]: E1013 09:06:49.730666 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b46cbc3-a4eb-4e31-8939-139e60849422" containerName="init" Oct 13 09:06:49 crc kubenswrapper[4685]: I1013 09:06:49.730674 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b46cbc3-a4eb-4e31-8939-139e60849422" containerName="init" Oct 13 09:06:49 crc kubenswrapper[4685]: E1013 09:06:49.730683 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfd5d590-0f0f-4de4-a373-ac9311796783" containerName="init" Oct 13 09:06:49 crc kubenswrapper[4685]: I1013 09:06:49.730691 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfd5d590-0f0f-4de4-a373-ac9311796783" containerName="init" Oct 13 09:06:49 crc kubenswrapper[4685]: E1013 09:06:49.730790 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b46cbc3-a4eb-4e31-8939-139e60849422" containerName="dnsmasq-dns" Oct 13 09:06:49 crc kubenswrapper[4685]: I1013 09:06:49.730798 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b46cbc3-a4eb-4e31-8939-139e60849422" containerName="dnsmasq-dns" Oct 13 09:06:49 crc kubenswrapper[4685]: I1013 09:06:49.731027 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfd5d590-0f0f-4de4-a373-ac9311796783" containerName="dnsmasq-dns" Oct 13 09:06:49 crc kubenswrapper[4685]: I1013 09:06:49.731054 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b46cbc3-a4eb-4e31-8939-139e60849422" containerName="dnsmasq-dns" Oct 13 09:06:49 crc kubenswrapper[4685]: I1013 09:06:49.732037 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc" Oct 13 09:06:49 crc kubenswrapper[4685]: I1013 09:06:49.735686 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 09:06:49 crc kubenswrapper[4685]: I1013 09:06:49.736215 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hcn4z" Oct 13 09:06:49 crc kubenswrapper[4685]: I1013 09:06:49.736253 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 09:06:49 crc kubenswrapper[4685]: I1013 09:06:49.737319 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 09:06:49 crc kubenswrapper[4685]: I1013 09:06:49.790531 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc"] Oct 13 09:06:49 crc kubenswrapper[4685]: I1013 09:06:49.892205 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc\" (UID: \"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc" Oct 13 09:06:49 crc kubenswrapper[4685]: I1013 09:06:49.892274 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc\" (UID: \"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc" Oct 13 09:06:49 crc kubenswrapper[4685]: I1013 09:06:49.892923 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc\" (UID: \"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc" Oct 13 09:06:49 crc kubenswrapper[4685]: I1013 09:06:49.892992 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfljh\" (UniqueName: \"kubernetes.io/projected/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-kube-api-access-dfljh\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc\" (UID: \"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc" Oct 13 09:06:49 crc kubenswrapper[4685]: I1013 09:06:49.994232 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc\" (UID: \"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc" Oct 13 09:06:49 crc kubenswrapper[4685]: I1013 09:06:49.994303 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc\" (UID: \"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc" Oct 13 09:06:49 crc kubenswrapper[4685]: I1013 09:06:49.994388 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc\" (UID: \"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc" Oct 13 09:06:49 crc kubenswrapper[4685]: I1013 09:06:49.994438 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfljh\" (UniqueName: \"kubernetes.io/projected/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-kube-api-access-dfljh\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc\" (UID: \"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc" Oct 13 09:06:50 crc kubenswrapper[4685]: I1013 09:06:50.002073 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc\" (UID: \"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc" Oct 13 09:06:50 crc kubenswrapper[4685]: I1013 09:06:50.010346 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc\" (UID: \"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc" Oct 13 09:06:50 crc kubenswrapper[4685]: I1013 09:06:50.027118 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfljh\" (UniqueName: \"kubernetes.io/projected/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-kube-api-access-dfljh\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc\" (UID: \"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc" Oct 13 09:06:50 crc kubenswrapper[4685]: I1013 09:06:50.028551 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc\" (UID: \"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc" Oct 13 09:06:50 crc kubenswrapper[4685]: I1013 09:06:50.051972 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc" Oct 13 09:06:50 crc kubenswrapper[4685]: I1013 09:06:50.899148 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc"] Oct 13 09:06:51 crc kubenswrapper[4685]: I1013 09:06:51.289205 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc" event={"ID":"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881","Type":"ContainerStarted","Data":"f1f451669dfff34af8e4332bd629e1ed173d208c9b2359fdab097a11f1441319"} Oct 13 09:06:53 crc kubenswrapper[4685]: I1013 09:06:53.627073 4685 scope.go:117] "RemoveContainer" containerID="692441bc08af17df5ba27504a2d82f321da33197cff11964acc63b131225301e" Oct 13 09:06:53 crc kubenswrapper[4685]: I1013 09:06:53.672548 4685 scope.go:117] "RemoveContainer" containerID="a8c575675b8a8240f9a9cce662be0dbf91b784093b2ec63f6e3d873f5d79efc2" Oct 13 09:06:56 crc kubenswrapper[4685]: I1013 09:06:56.339626 4685 generic.go:334] "Generic (PLEG): container finished" podID="b001d17a-1aea-44ba-86c5-ba6b312156c1" containerID="6f8ac1889cdb78c30c1167f41924fc8421f395f14a8be5ba49377fbd32fe2472" exitCode=1 Oct 13 09:06:56 crc kubenswrapper[4685]: I1013 09:06:56.339810 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" event={"ID":"b001d17a-1aea-44ba-86c5-ba6b312156c1","Type":"ContainerDied","Data":"6f8ac1889cdb78c30c1167f41924fc8421f395f14a8be5ba49377fbd32fe2472"} Oct 13 09:06:56 crc kubenswrapper[4685]: I1013 09:06:56.340228 4685 scope.go:117] "RemoveContainer" containerID="c44b3a507602babf650e62c8b2c5a5ac1d0f13d11239f82857414365cf156ea7" Oct 13 09:06:56 crc kubenswrapper[4685]: I1013 09:06:56.341343 4685 scope.go:117] "RemoveContainer" containerID="6f8ac1889cdb78c30c1167f41924fc8421f395f14a8be5ba49377fbd32fe2472" Oct 13 09:06:56 crc kubenswrapper[4685]: E1013 09:06:56.341665 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:06:58 crc kubenswrapper[4685]: I1013 09:06:58.309614 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 13 09:06:58 crc kubenswrapper[4685]: I1013 09:06:58.580073 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 13 09:07:02 crc kubenswrapper[4685]: I1013 09:07:02.419771 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc" event={"ID":"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881","Type":"ContainerStarted","Data":"b00ea06cc16d3823bf0fdeebecc7c93081858f293c908689c1ded2e635d495af"} Oct 13 09:07:02 crc kubenswrapper[4685]: I1013 09:07:02.444736 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc" podStartSLOduration=3.167698088 podStartE2EDuration="13.444717662s" podCreationTimestamp="2025-10-13 09:06:49 +0000 UTC" firstStartedPulling="2025-10-13 09:06:50.90273096 +0000 UTC m=+1336.050606711" lastFinishedPulling="2025-10-13 09:07:01.179750524 +0000 UTC m=+1346.327626285" observedRunningTime="2025-10-13 09:07:02.433418685 +0000 UTC m=+1347.581294456" watchObservedRunningTime="2025-10-13 09:07:02.444717662 +0000 UTC m=+1347.592593433" Oct 13 09:07:03 crc kubenswrapper[4685]: I1013 09:07:03.442683 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:07:03 crc kubenswrapper[4685]: I1013 09:07:03.442737 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:07:03 crc kubenswrapper[4685]: I1013 09:07:03.443395 4685 scope.go:117] "RemoveContainer" containerID="6f8ac1889cdb78c30c1167f41924fc8421f395f14a8be5ba49377fbd32fe2472" Oct 13 09:07:03 crc kubenswrapper[4685]: E1013 09:07:03.443636 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:07:18 crc kubenswrapper[4685]: I1013 09:07:18.502992 4685 scope.go:117] "RemoveContainer" containerID="6f8ac1889cdb78c30c1167f41924fc8421f395f14a8be5ba49377fbd32fe2472" Oct 13 09:07:19 crc kubenswrapper[4685]: I1013 09:07:19.673042 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" event={"ID":"b001d17a-1aea-44ba-86c5-ba6b312156c1","Type":"ContainerStarted","Data":"cedf68f34af27ba6b0e7b6a3ebc8c368caa3b020eacec7a71ecf4f542d94655f"} Oct 13 09:07:19 crc kubenswrapper[4685]: I1013 09:07:19.674869 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:07:23 crc kubenswrapper[4685]: I1013 09:07:23.446744 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:07:24 crc kubenswrapper[4685]: I1013 09:07:24.718454 4685 generic.go:334] "Generic (PLEG): container finished" podID="81fecbf5-ba69-4c64-b5c1-ae9f5e07d881" containerID="b00ea06cc16d3823bf0fdeebecc7c93081858f293c908689c1ded2e635d495af" exitCode=0 Oct 13 09:07:24 crc kubenswrapper[4685]: I1013 09:07:24.718501 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc" event={"ID":"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881","Type":"ContainerDied","Data":"b00ea06cc16d3823bf0fdeebecc7c93081858f293c908689c1ded2e635d495af"} Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.225471 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc" Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.306464 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfljh\" (UniqueName: \"kubernetes.io/projected/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-kube-api-access-dfljh\") pod \"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881\" (UID: \"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881\") " Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.306940 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-repo-setup-combined-ca-bundle\") pod \"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881\" (UID: \"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881\") " Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.307128 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-inventory\") pod \"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881\" (UID: \"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881\") " Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.307283 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-ssh-key\") pod \"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881\" (UID: \"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881\") " Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.312829 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-kube-api-access-dfljh" (OuterVolumeSpecName: "kube-api-access-dfljh") pod "81fecbf5-ba69-4c64-b5c1-ae9f5e07d881" (UID: "81fecbf5-ba69-4c64-b5c1-ae9f5e07d881"). InnerVolumeSpecName "kube-api-access-dfljh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.313258 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "81fecbf5-ba69-4c64-b5c1-ae9f5e07d881" (UID: "81fecbf5-ba69-4c64-b5c1-ae9f5e07d881"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.349756 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "81fecbf5-ba69-4c64-b5c1-ae9f5e07d881" (UID: "81fecbf5-ba69-4c64-b5c1-ae9f5e07d881"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.356135 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-inventory" (OuterVolumeSpecName: "inventory") pod "81fecbf5-ba69-4c64-b5c1-ae9f5e07d881" (UID: "81fecbf5-ba69-4c64-b5c1-ae9f5e07d881"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.409964 4685 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.410000 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.410010 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.410019 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfljh\" (UniqueName: \"kubernetes.io/projected/81fecbf5-ba69-4c64-b5c1-ae9f5e07d881-kube-api-access-dfljh\") on node \"crc\" DevicePath \"\"" Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.741046 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc" event={"ID":"81fecbf5-ba69-4c64-b5c1-ae9f5e07d881","Type":"ContainerDied","Data":"f1f451669dfff34af8e4332bd629e1ed173d208c9b2359fdab097a11f1441319"} Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.741100 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1f451669dfff34af8e4332bd629e1ed173d208c9b2359fdab097a11f1441319" Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.741165 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc" Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.832424 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-gbjgj"] Oct 13 09:07:26 crc kubenswrapper[4685]: E1013 09:07:26.832813 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81fecbf5-ba69-4c64-b5c1-ae9f5e07d881" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.832834 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="81fecbf5-ba69-4c64-b5c1-ae9f5e07d881" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.833044 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="81fecbf5-ba69-4c64-b5c1-ae9f5e07d881" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.833632 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gbjgj" Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.835527 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.835602 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.838780 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.845022 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hcn4z" Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.852087 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-gbjgj"] Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.917972 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc857d68-fe9c-4e34-86c7-e26aca5432e6-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gbjgj\" (UID: \"cc857d68-fe9c-4e34-86c7-e26aca5432e6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gbjgj" Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.918028 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcv5l\" (UniqueName: \"kubernetes.io/projected/cc857d68-fe9c-4e34-86c7-e26aca5432e6-kube-api-access-vcv5l\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gbjgj\" (UID: \"cc857d68-fe9c-4e34-86c7-e26aca5432e6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gbjgj" Oct 13 09:07:26 crc kubenswrapper[4685]: I1013 09:07:26.918077 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc857d68-fe9c-4e34-86c7-e26aca5432e6-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gbjgj\" (UID: \"cc857d68-fe9c-4e34-86c7-e26aca5432e6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gbjgj" Oct 13 09:07:27 crc kubenswrapper[4685]: I1013 09:07:27.020195 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc857d68-fe9c-4e34-86c7-e26aca5432e6-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gbjgj\" (UID: \"cc857d68-fe9c-4e34-86c7-e26aca5432e6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gbjgj" Oct 13 09:07:27 crc kubenswrapper[4685]: I1013 09:07:27.020243 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcv5l\" (UniqueName: \"kubernetes.io/projected/cc857d68-fe9c-4e34-86c7-e26aca5432e6-kube-api-access-vcv5l\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gbjgj\" (UID: \"cc857d68-fe9c-4e34-86c7-e26aca5432e6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gbjgj" Oct 13 09:07:27 crc kubenswrapper[4685]: I1013 09:07:27.020280 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc857d68-fe9c-4e34-86c7-e26aca5432e6-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gbjgj\" (UID: \"cc857d68-fe9c-4e34-86c7-e26aca5432e6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gbjgj" Oct 13 09:07:27 crc kubenswrapper[4685]: I1013 09:07:27.025693 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc857d68-fe9c-4e34-86c7-e26aca5432e6-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gbjgj\" (UID: \"cc857d68-fe9c-4e34-86c7-e26aca5432e6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gbjgj" Oct 13 09:07:27 crc kubenswrapper[4685]: I1013 09:07:27.037946 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc857d68-fe9c-4e34-86c7-e26aca5432e6-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gbjgj\" (UID: \"cc857d68-fe9c-4e34-86c7-e26aca5432e6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gbjgj" Oct 13 09:07:27 crc kubenswrapper[4685]: I1013 09:07:27.042852 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcv5l\" (UniqueName: \"kubernetes.io/projected/cc857d68-fe9c-4e34-86c7-e26aca5432e6-kube-api-access-vcv5l\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gbjgj\" (UID: \"cc857d68-fe9c-4e34-86c7-e26aca5432e6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gbjgj" Oct 13 09:07:27 crc kubenswrapper[4685]: I1013 09:07:27.156094 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gbjgj" Oct 13 09:07:27 crc kubenswrapper[4685]: I1013 09:07:27.670345 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-gbjgj"] Oct 13 09:07:27 crc kubenswrapper[4685]: I1013 09:07:27.752049 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gbjgj" event={"ID":"cc857d68-fe9c-4e34-86c7-e26aca5432e6","Type":"ContainerStarted","Data":"fe3a90952dfc319b7136bf455c0ff1583d5db139e2903eea8c20a5f0a0efbbb5"} Oct 13 09:07:28 crc kubenswrapper[4685]: I1013 09:07:28.762853 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gbjgj" event={"ID":"cc857d68-fe9c-4e34-86c7-e26aca5432e6","Type":"ContainerStarted","Data":"fced1be428e0ce6c96f232c96fffc904f7fdfca52efca933b609df1ec4c0040c"} Oct 13 09:07:28 crc kubenswrapper[4685]: I1013 09:07:28.788796 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gbjgj" podStartSLOduration=2.222304192 podStartE2EDuration="2.788769151s" podCreationTimestamp="2025-10-13 09:07:26 +0000 UTC" firstStartedPulling="2025-10-13 09:07:27.678677557 +0000 UTC m=+1372.826553318" lastFinishedPulling="2025-10-13 09:07:28.245142526 +0000 UTC m=+1373.393018277" observedRunningTime="2025-10-13 09:07:28.779701524 +0000 UTC m=+1373.927577285" watchObservedRunningTime="2025-10-13 09:07:28.788769151 +0000 UTC m=+1373.936644912" Oct 13 09:07:31 crc kubenswrapper[4685]: E1013 09:07:31.278180 4685 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc857d68_fe9c_4e34_86c7_e26aca5432e6.slice/crio-conmon-fced1be428e0ce6c96f232c96fffc904f7fdfca52efca933b609df1ec4c0040c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc857d68_fe9c_4e34_86c7_e26aca5432e6.slice/crio-fced1be428e0ce6c96f232c96fffc904f7fdfca52efca933b609df1ec4c0040c.scope\": RecentStats: unable to find data in memory cache]" Oct 13 09:07:31 crc kubenswrapper[4685]: I1013 09:07:31.813809 4685 generic.go:334] "Generic (PLEG): container finished" podID="cc857d68-fe9c-4e34-86c7-e26aca5432e6" containerID="fced1be428e0ce6c96f232c96fffc904f7fdfca52efca933b609df1ec4c0040c" exitCode=0 Oct 13 09:07:31 crc kubenswrapper[4685]: I1013 09:07:31.813878 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gbjgj" event={"ID":"cc857d68-fe9c-4e34-86c7-e26aca5432e6","Type":"ContainerDied","Data":"fced1be428e0ce6c96f232c96fffc904f7fdfca52efca933b609df1ec4c0040c"} Oct 13 09:07:33 crc kubenswrapper[4685]: I1013 09:07:33.313844 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gbjgj" Oct 13 09:07:33 crc kubenswrapper[4685]: I1013 09:07:33.447254 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc857d68-fe9c-4e34-86c7-e26aca5432e6-ssh-key\") pod \"cc857d68-fe9c-4e34-86c7-e26aca5432e6\" (UID: \"cc857d68-fe9c-4e34-86c7-e26aca5432e6\") " Oct 13 09:07:33 crc kubenswrapper[4685]: I1013 09:07:33.447421 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcv5l\" (UniqueName: \"kubernetes.io/projected/cc857d68-fe9c-4e34-86c7-e26aca5432e6-kube-api-access-vcv5l\") pod \"cc857d68-fe9c-4e34-86c7-e26aca5432e6\" (UID: \"cc857d68-fe9c-4e34-86c7-e26aca5432e6\") " Oct 13 09:07:33 crc kubenswrapper[4685]: I1013 09:07:33.447575 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc857d68-fe9c-4e34-86c7-e26aca5432e6-inventory\") pod \"cc857d68-fe9c-4e34-86c7-e26aca5432e6\" (UID: \"cc857d68-fe9c-4e34-86c7-e26aca5432e6\") " Oct 13 09:07:33 crc kubenswrapper[4685]: I1013 09:07:33.453304 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc857d68-fe9c-4e34-86c7-e26aca5432e6-kube-api-access-vcv5l" (OuterVolumeSpecName: "kube-api-access-vcv5l") pod "cc857d68-fe9c-4e34-86c7-e26aca5432e6" (UID: "cc857d68-fe9c-4e34-86c7-e26aca5432e6"). InnerVolumeSpecName "kube-api-access-vcv5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:07:33 crc kubenswrapper[4685]: I1013 09:07:33.478014 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc857d68-fe9c-4e34-86c7-e26aca5432e6-inventory" (OuterVolumeSpecName: "inventory") pod "cc857d68-fe9c-4e34-86c7-e26aca5432e6" (UID: "cc857d68-fe9c-4e34-86c7-e26aca5432e6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:07:33 crc kubenswrapper[4685]: I1013 09:07:33.480681 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc857d68-fe9c-4e34-86c7-e26aca5432e6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cc857d68-fe9c-4e34-86c7-e26aca5432e6" (UID: "cc857d68-fe9c-4e34-86c7-e26aca5432e6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:07:33 crc kubenswrapper[4685]: I1013 09:07:33.549790 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcv5l\" (UniqueName: \"kubernetes.io/projected/cc857d68-fe9c-4e34-86c7-e26aca5432e6-kube-api-access-vcv5l\") on node \"crc\" DevicePath \"\"" Oct 13 09:07:33 crc kubenswrapper[4685]: I1013 09:07:33.549827 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc857d68-fe9c-4e34-86c7-e26aca5432e6-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 09:07:33 crc kubenswrapper[4685]: I1013 09:07:33.549840 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc857d68-fe9c-4e34-86c7-e26aca5432e6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 09:07:33 crc kubenswrapper[4685]: I1013 09:07:33.837396 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gbjgj" Oct 13 09:07:33 crc kubenswrapper[4685]: I1013 09:07:33.837436 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gbjgj" event={"ID":"cc857d68-fe9c-4e34-86c7-e26aca5432e6","Type":"ContainerDied","Data":"fe3a90952dfc319b7136bf455c0ff1583d5db139e2903eea8c20a5f0a0efbbb5"} Oct 13 09:07:33 crc kubenswrapper[4685]: I1013 09:07:33.837487 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe3a90952dfc319b7136bf455c0ff1583d5db139e2903eea8c20a5f0a0efbbb5" Oct 13 09:07:33 crc kubenswrapper[4685]: I1013 09:07:33.924478 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl"] Oct 13 09:07:33 crc kubenswrapper[4685]: E1013 09:07:33.925043 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc857d68-fe9c-4e34-86c7-e26aca5432e6" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 13 09:07:33 crc kubenswrapper[4685]: I1013 09:07:33.925064 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc857d68-fe9c-4e34-86c7-e26aca5432e6" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 13 09:07:33 crc kubenswrapper[4685]: I1013 09:07:33.925276 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc857d68-fe9c-4e34-86c7-e26aca5432e6" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 13 09:07:33 crc kubenswrapper[4685]: I1013 09:07:33.926053 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl" Oct 13 09:07:33 crc kubenswrapper[4685]: I1013 09:07:33.930696 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 09:07:33 crc kubenswrapper[4685]: I1013 09:07:33.930976 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 09:07:33 crc kubenswrapper[4685]: I1013 09:07:33.931139 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hcn4z" Oct 13 09:07:33 crc kubenswrapper[4685]: I1013 09:07:33.931175 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 09:07:33 crc kubenswrapper[4685]: I1013 09:07:33.937046 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl"] Oct 13 09:07:34 crc kubenswrapper[4685]: I1013 09:07:34.063712 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxgn8\" (UniqueName: \"kubernetes.io/projected/5056abd4-630c-49fe-9ddf-1294f0f6b55a-kube-api-access-qxgn8\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl\" (UID: \"5056abd4-630c-49fe-9ddf-1294f0f6b55a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl" Oct 13 09:07:34 crc kubenswrapper[4685]: I1013 09:07:34.063948 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5056abd4-630c-49fe-9ddf-1294f0f6b55a-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl\" (UID: \"5056abd4-630c-49fe-9ddf-1294f0f6b55a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl" Oct 13 09:07:34 crc kubenswrapper[4685]: I1013 09:07:34.064080 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5056abd4-630c-49fe-9ddf-1294f0f6b55a-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl\" (UID: \"5056abd4-630c-49fe-9ddf-1294f0f6b55a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl" Oct 13 09:07:34 crc kubenswrapper[4685]: I1013 09:07:34.064267 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5056abd4-630c-49fe-9ddf-1294f0f6b55a-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl\" (UID: \"5056abd4-630c-49fe-9ddf-1294f0f6b55a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl" Oct 13 09:07:34 crc kubenswrapper[4685]: I1013 09:07:34.166269 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5056abd4-630c-49fe-9ddf-1294f0f6b55a-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl\" (UID: \"5056abd4-630c-49fe-9ddf-1294f0f6b55a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl" Oct 13 09:07:34 crc kubenswrapper[4685]: I1013 09:07:34.166658 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxgn8\" (UniqueName: \"kubernetes.io/projected/5056abd4-630c-49fe-9ddf-1294f0f6b55a-kube-api-access-qxgn8\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl\" (UID: \"5056abd4-630c-49fe-9ddf-1294f0f6b55a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl" Oct 13 09:07:34 crc kubenswrapper[4685]: I1013 09:07:34.166829 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5056abd4-630c-49fe-9ddf-1294f0f6b55a-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl\" (UID: \"5056abd4-630c-49fe-9ddf-1294f0f6b55a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl" Oct 13 09:07:34 crc kubenswrapper[4685]: I1013 09:07:34.167088 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5056abd4-630c-49fe-9ddf-1294f0f6b55a-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl\" (UID: \"5056abd4-630c-49fe-9ddf-1294f0f6b55a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl" Oct 13 09:07:34 crc kubenswrapper[4685]: I1013 09:07:34.170585 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5056abd4-630c-49fe-9ddf-1294f0f6b55a-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl\" (UID: \"5056abd4-630c-49fe-9ddf-1294f0f6b55a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl" Oct 13 09:07:34 crc kubenswrapper[4685]: I1013 09:07:34.170818 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5056abd4-630c-49fe-9ddf-1294f0f6b55a-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl\" (UID: \"5056abd4-630c-49fe-9ddf-1294f0f6b55a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl" Oct 13 09:07:34 crc kubenswrapper[4685]: I1013 09:07:34.171604 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5056abd4-630c-49fe-9ddf-1294f0f6b55a-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl\" (UID: \"5056abd4-630c-49fe-9ddf-1294f0f6b55a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl" Oct 13 09:07:34 crc kubenswrapper[4685]: I1013 09:07:34.192153 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxgn8\" (UniqueName: \"kubernetes.io/projected/5056abd4-630c-49fe-9ddf-1294f0f6b55a-kube-api-access-qxgn8\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl\" (UID: \"5056abd4-630c-49fe-9ddf-1294f0f6b55a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl" Oct 13 09:07:34 crc kubenswrapper[4685]: I1013 09:07:34.254563 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl" Oct 13 09:07:34 crc kubenswrapper[4685]: I1013 09:07:34.794876 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl"] Oct 13 09:07:34 crc kubenswrapper[4685]: W1013 09:07:34.803211 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5056abd4_630c_49fe_9ddf_1294f0f6b55a.slice/crio-9ae9f3547a9800eccb87ab49c090b7f0ccecaac0ffa1100ef286bfe7569b52ff WatchSource:0}: Error finding container 9ae9f3547a9800eccb87ab49c090b7f0ccecaac0ffa1100ef286bfe7569b52ff: Status 404 returned error can't find the container with id 9ae9f3547a9800eccb87ab49c090b7f0ccecaac0ffa1100ef286bfe7569b52ff Oct 13 09:07:34 crc kubenswrapper[4685]: I1013 09:07:34.848722 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl" event={"ID":"5056abd4-630c-49fe-9ddf-1294f0f6b55a","Type":"ContainerStarted","Data":"9ae9f3547a9800eccb87ab49c090b7f0ccecaac0ffa1100ef286bfe7569b52ff"} Oct 13 09:07:35 crc kubenswrapper[4685]: I1013 09:07:35.860470 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl" event={"ID":"5056abd4-630c-49fe-9ddf-1294f0f6b55a","Type":"ContainerStarted","Data":"6131eec83299a39e69b777f0edd29636b60749847c97c68d2af25429890bdd58"} Oct 13 09:07:35 crc kubenswrapper[4685]: I1013 09:07:35.883485 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl" podStartSLOduration=2.502754795 podStartE2EDuration="2.883463916s" podCreationTimestamp="2025-10-13 09:07:33 +0000 UTC" firstStartedPulling="2025-10-13 09:07:34.805824707 +0000 UTC m=+1379.953700468" lastFinishedPulling="2025-10-13 09:07:35.186533818 +0000 UTC m=+1380.334409589" observedRunningTime="2025-10-13 09:07:35.876344623 +0000 UTC m=+1381.024220424" watchObservedRunningTime="2025-10-13 09:07:35.883463916 +0000 UTC m=+1381.031339677" Oct 13 09:07:52 crc kubenswrapper[4685]: I1013 09:07:52.032103 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xl64r"] Oct 13 09:07:52 crc kubenswrapper[4685]: I1013 09:07:52.035115 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xl64r" Oct 13 09:07:52 crc kubenswrapper[4685]: I1013 09:07:52.072156 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xl64r"] Oct 13 09:07:52 crc kubenswrapper[4685]: I1013 09:07:52.128127 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a4fa29-ba61-4eb0-94b1-d4d30daa135a-catalog-content\") pod \"community-operators-xl64r\" (UID: \"d4a4fa29-ba61-4eb0-94b1-d4d30daa135a\") " pod="openshift-marketplace/community-operators-xl64r" Oct 13 09:07:52 crc kubenswrapper[4685]: I1013 09:07:52.128292 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hn5f\" (UniqueName: \"kubernetes.io/projected/d4a4fa29-ba61-4eb0-94b1-d4d30daa135a-kube-api-access-7hn5f\") pod \"community-operators-xl64r\" (UID: \"d4a4fa29-ba61-4eb0-94b1-d4d30daa135a\") " pod="openshift-marketplace/community-operators-xl64r" Oct 13 09:07:52 crc kubenswrapper[4685]: I1013 09:07:52.128338 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a4fa29-ba61-4eb0-94b1-d4d30daa135a-utilities\") pod \"community-operators-xl64r\" (UID: \"d4a4fa29-ba61-4eb0-94b1-d4d30daa135a\") " pod="openshift-marketplace/community-operators-xl64r" Oct 13 09:07:52 crc kubenswrapper[4685]: I1013 09:07:52.230573 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hn5f\" (UniqueName: \"kubernetes.io/projected/d4a4fa29-ba61-4eb0-94b1-d4d30daa135a-kube-api-access-7hn5f\") pod \"community-operators-xl64r\" (UID: \"d4a4fa29-ba61-4eb0-94b1-d4d30daa135a\") " pod="openshift-marketplace/community-operators-xl64r" Oct 13 09:07:52 crc kubenswrapper[4685]: I1013 09:07:52.230637 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a4fa29-ba61-4eb0-94b1-d4d30daa135a-utilities\") pod \"community-operators-xl64r\" (UID: \"d4a4fa29-ba61-4eb0-94b1-d4d30daa135a\") " pod="openshift-marketplace/community-operators-xl64r" Oct 13 09:07:52 crc kubenswrapper[4685]: I1013 09:07:52.230772 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a4fa29-ba61-4eb0-94b1-d4d30daa135a-catalog-content\") pod \"community-operators-xl64r\" (UID: \"d4a4fa29-ba61-4eb0-94b1-d4d30daa135a\") " pod="openshift-marketplace/community-operators-xl64r" Oct 13 09:07:52 crc kubenswrapper[4685]: I1013 09:07:52.231269 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a4fa29-ba61-4eb0-94b1-d4d30daa135a-utilities\") pod \"community-operators-xl64r\" (UID: \"d4a4fa29-ba61-4eb0-94b1-d4d30daa135a\") " pod="openshift-marketplace/community-operators-xl64r" Oct 13 09:07:52 crc kubenswrapper[4685]: I1013 09:07:52.231335 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a4fa29-ba61-4eb0-94b1-d4d30daa135a-catalog-content\") pod \"community-operators-xl64r\" (UID: \"d4a4fa29-ba61-4eb0-94b1-d4d30daa135a\") " pod="openshift-marketplace/community-operators-xl64r" Oct 13 09:07:52 crc kubenswrapper[4685]: I1013 09:07:52.251701 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hn5f\" (UniqueName: \"kubernetes.io/projected/d4a4fa29-ba61-4eb0-94b1-d4d30daa135a-kube-api-access-7hn5f\") pod \"community-operators-xl64r\" (UID: \"d4a4fa29-ba61-4eb0-94b1-d4d30daa135a\") " pod="openshift-marketplace/community-operators-xl64r" Oct 13 09:07:52 crc kubenswrapper[4685]: I1013 09:07:52.380216 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xl64r" Oct 13 09:07:52 crc kubenswrapper[4685]: I1013 09:07:52.965669 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xl64r"] Oct 13 09:07:52 crc kubenswrapper[4685]: W1013 09:07:52.968867 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4a4fa29_ba61_4eb0_94b1_d4d30daa135a.slice/crio-30596a907873e5ea0e4a92c5623a40e6d978c655023dca684a01f684bf7ccb93 WatchSource:0}: Error finding container 30596a907873e5ea0e4a92c5623a40e6d978c655023dca684a01f684bf7ccb93: Status 404 returned error can't find the container with id 30596a907873e5ea0e4a92c5623a40e6d978c655023dca684a01f684bf7ccb93 Oct 13 09:07:53 crc kubenswrapper[4685]: I1013 09:07:53.086853 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xl64r" event={"ID":"d4a4fa29-ba61-4eb0-94b1-d4d30daa135a","Type":"ContainerStarted","Data":"30596a907873e5ea0e4a92c5623a40e6d978c655023dca684a01f684bf7ccb93"} Oct 13 09:07:53 crc kubenswrapper[4685]: I1013 09:07:53.787818 4685 scope.go:117] "RemoveContainer" containerID="b645002af7b6e1599aca65be6d1ea012273493a144aa47de695155e939648e51" Oct 13 09:07:53 crc kubenswrapper[4685]: I1013 09:07:53.811926 4685 scope.go:117] "RemoveContainer" containerID="0df2a8d7eb43f1a38cf4e1b537678a2c97537c359396ac708b4e4ca59a9b58cf" Oct 13 09:07:53 crc kubenswrapper[4685]: I1013 09:07:53.842814 4685 scope.go:117] "RemoveContainer" containerID="09eb66b97da1f79efa9078a0e2afd34299176faadf74af83bf9b0b03d570003f" Oct 13 09:07:53 crc kubenswrapper[4685]: I1013 09:07:53.893044 4685 scope.go:117] "RemoveContainer" containerID="99516221fed8e294d62a6cb32d71ed431230ed3bc0d03203306b0083cec2be29" Oct 13 09:07:53 crc kubenswrapper[4685]: I1013 09:07:53.934212 4685 scope.go:117] "RemoveContainer" containerID="a5aee9e8730b63472aa83483317267b5ae11e7ad15304edd44922a53a600a0cb" Oct 13 09:07:53 crc kubenswrapper[4685]: I1013 09:07:53.957051 4685 scope.go:117] "RemoveContainer" containerID="2abbed46b374258b1c20646f9e96b0579634e4fd1f5cfdf374dbbef34bc7d198" Oct 13 09:07:54 crc kubenswrapper[4685]: I1013 09:07:54.011124 4685 scope.go:117] "RemoveContainer" containerID="6faa94ee30712a6d42db9f6e7ed17637643bdf663c088d057871184b5370d738" Oct 13 09:07:54 crc kubenswrapper[4685]: I1013 09:07:54.109067 4685 generic.go:334] "Generic (PLEG): container finished" podID="d4a4fa29-ba61-4eb0-94b1-d4d30daa135a" containerID="e8ccecbf9cae068f16592f61323b69de10a9f7f1116a911467be425f35bedd44" exitCode=0 Oct 13 09:07:54 crc kubenswrapper[4685]: I1013 09:07:54.109121 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xl64r" event={"ID":"d4a4fa29-ba61-4eb0-94b1-d4d30daa135a","Type":"ContainerDied","Data":"e8ccecbf9cae068f16592f61323b69de10a9f7f1116a911467be425f35bedd44"} Oct 13 09:07:56 crc kubenswrapper[4685]: I1013 09:07:56.135983 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xl64r" event={"ID":"d4a4fa29-ba61-4eb0-94b1-d4d30daa135a","Type":"ContainerStarted","Data":"afb073b510b0d675317e4358488138193f1aeb7708e1f3b8b34069dd785cdddf"} Oct 13 09:07:57 crc kubenswrapper[4685]: I1013 09:07:57.145827 4685 generic.go:334] "Generic (PLEG): container finished" podID="d4a4fa29-ba61-4eb0-94b1-d4d30daa135a" containerID="afb073b510b0d675317e4358488138193f1aeb7708e1f3b8b34069dd785cdddf" exitCode=0 Oct 13 09:07:57 crc kubenswrapper[4685]: I1013 09:07:57.145948 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xl64r" event={"ID":"d4a4fa29-ba61-4eb0-94b1-d4d30daa135a","Type":"ContainerDied","Data":"afb073b510b0d675317e4358488138193f1aeb7708e1f3b8b34069dd785cdddf"} Oct 13 09:07:58 crc kubenswrapper[4685]: I1013 09:07:58.160203 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xl64r" event={"ID":"d4a4fa29-ba61-4eb0-94b1-d4d30daa135a","Type":"ContainerStarted","Data":"c84b91750d1f1ce6210d0fcb25296eac5754f23401474dfebd500c1101716adb"} Oct 13 09:07:58 crc kubenswrapper[4685]: I1013 09:07:58.181058 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xl64r" podStartSLOduration=2.683575039 podStartE2EDuration="6.18103823s" podCreationTimestamp="2025-10-13 09:07:52 +0000 UTC" firstStartedPulling="2025-10-13 09:07:54.131294728 +0000 UTC m=+1399.279170489" lastFinishedPulling="2025-10-13 09:07:57.628757919 +0000 UTC m=+1402.776633680" observedRunningTime="2025-10-13 09:07:58.176302651 +0000 UTC m=+1403.324178442" watchObservedRunningTime="2025-10-13 09:07:58.18103823 +0000 UTC m=+1403.328914001" Oct 13 09:08:01 crc kubenswrapper[4685]: I1013 09:08:01.033087 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8qfp7"] Oct 13 09:08:01 crc kubenswrapper[4685]: I1013 09:08:01.036824 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8qfp7" Oct 13 09:08:01 crc kubenswrapper[4685]: I1013 09:08:01.047082 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8qfp7"] Oct 13 09:08:01 crc kubenswrapper[4685]: I1013 09:08:01.114583 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rd62w\" (UniqueName: \"kubernetes.io/projected/9b1f2f60-a981-432b-a7cb-f84d1c4a80c3-kube-api-access-rd62w\") pod \"certified-operators-8qfp7\" (UID: \"9b1f2f60-a981-432b-a7cb-f84d1c4a80c3\") " pod="openshift-marketplace/certified-operators-8qfp7" Oct 13 09:08:01 crc kubenswrapper[4685]: I1013 09:08:01.114884 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b1f2f60-a981-432b-a7cb-f84d1c4a80c3-utilities\") pod \"certified-operators-8qfp7\" (UID: \"9b1f2f60-a981-432b-a7cb-f84d1c4a80c3\") " pod="openshift-marketplace/certified-operators-8qfp7" Oct 13 09:08:01 crc kubenswrapper[4685]: I1013 09:08:01.115296 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b1f2f60-a981-432b-a7cb-f84d1c4a80c3-catalog-content\") pod \"certified-operators-8qfp7\" (UID: \"9b1f2f60-a981-432b-a7cb-f84d1c4a80c3\") " pod="openshift-marketplace/certified-operators-8qfp7" Oct 13 09:08:01 crc kubenswrapper[4685]: I1013 09:08:01.217277 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b1f2f60-a981-432b-a7cb-f84d1c4a80c3-catalog-content\") pod \"certified-operators-8qfp7\" (UID: \"9b1f2f60-a981-432b-a7cb-f84d1c4a80c3\") " pod="openshift-marketplace/certified-operators-8qfp7" Oct 13 09:08:01 crc kubenswrapper[4685]: I1013 09:08:01.217377 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rd62w\" (UniqueName: \"kubernetes.io/projected/9b1f2f60-a981-432b-a7cb-f84d1c4a80c3-kube-api-access-rd62w\") pod \"certified-operators-8qfp7\" (UID: \"9b1f2f60-a981-432b-a7cb-f84d1c4a80c3\") " pod="openshift-marketplace/certified-operators-8qfp7" Oct 13 09:08:01 crc kubenswrapper[4685]: I1013 09:08:01.217401 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b1f2f60-a981-432b-a7cb-f84d1c4a80c3-utilities\") pod \"certified-operators-8qfp7\" (UID: \"9b1f2f60-a981-432b-a7cb-f84d1c4a80c3\") " pod="openshift-marketplace/certified-operators-8qfp7" Oct 13 09:08:01 crc kubenswrapper[4685]: I1013 09:08:01.217870 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b1f2f60-a981-432b-a7cb-f84d1c4a80c3-utilities\") pod \"certified-operators-8qfp7\" (UID: \"9b1f2f60-a981-432b-a7cb-f84d1c4a80c3\") " pod="openshift-marketplace/certified-operators-8qfp7" Oct 13 09:08:01 crc kubenswrapper[4685]: I1013 09:08:01.218134 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b1f2f60-a981-432b-a7cb-f84d1c4a80c3-catalog-content\") pod \"certified-operators-8qfp7\" (UID: \"9b1f2f60-a981-432b-a7cb-f84d1c4a80c3\") " pod="openshift-marketplace/certified-operators-8qfp7" Oct 13 09:08:01 crc kubenswrapper[4685]: I1013 09:08:01.246650 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rd62w\" (UniqueName: \"kubernetes.io/projected/9b1f2f60-a981-432b-a7cb-f84d1c4a80c3-kube-api-access-rd62w\") pod \"certified-operators-8qfp7\" (UID: \"9b1f2f60-a981-432b-a7cb-f84d1c4a80c3\") " pod="openshift-marketplace/certified-operators-8qfp7" Oct 13 09:08:01 crc kubenswrapper[4685]: I1013 09:08:01.368493 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8qfp7" Oct 13 09:08:01 crc kubenswrapper[4685]: I1013 09:08:01.871382 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8qfp7"] Oct 13 09:08:01 crc kubenswrapper[4685]: W1013 09:08:01.873636 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b1f2f60_a981_432b_a7cb_f84d1c4a80c3.slice/crio-98d114306d773ede677073783ae9eff71840bd644145ef1b547cd8cfbb818148 WatchSource:0}: Error finding container 98d114306d773ede677073783ae9eff71840bd644145ef1b547cd8cfbb818148: Status 404 returned error can't find the container with id 98d114306d773ede677073783ae9eff71840bd644145ef1b547cd8cfbb818148 Oct 13 09:08:02 crc kubenswrapper[4685]: I1013 09:08:02.213020 4685 generic.go:334] "Generic (PLEG): container finished" podID="9b1f2f60-a981-432b-a7cb-f84d1c4a80c3" containerID="46718462dbf280e9b5afb8268be15a0166ae6e020be5991ccd55ef8f449cf9c6" exitCode=0 Oct 13 09:08:02 crc kubenswrapper[4685]: I1013 09:08:02.213065 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qfp7" event={"ID":"9b1f2f60-a981-432b-a7cb-f84d1c4a80c3","Type":"ContainerDied","Data":"46718462dbf280e9b5afb8268be15a0166ae6e020be5991ccd55ef8f449cf9c6"} Oct 13 09:08:02 crc kubenswrapper[4685]: I1013 09:08:02.213091 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qfp7" event={"ID":"9b1f2f60-a981-432b-a7cb-f84d1c4a80c3","Type":"ContainerStarted","Data":"98d114306d773ede677073783ae9eff71840bd644145ef1b547cd8cfbb818148"} Oct 13 09:08:02 crc kubenswrapper[4685]: I1013 09:08:02.381779 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xl64r" Oct 13 09:08:02 crc kubenswrapper[4685]: I1013 09:08:02.382147 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xl64r" Oct 13 09:08:03 crc kubenswrapper[4685]: I1013 09:08:03.430492 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-xl64r" podUID="d4a4fa29-ba61-4eb0-94b1-d4d30daa135a" containerName="registry-server" probeResult="failure" output=< Oct 13 09:08:03 crc kubenswrapper[4685]: timeout: failed to connect service ":50051" within 1s Oct 13 09:08:03 crc kubenswrapper[4685]: > Oct 13 09:08:04 crc kubenswrapper[4685]: I1013 09:08:04.233059 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qfp7" event={"ID":"9b1f2f60-a981-432b-a7cb-f84d1c4a80c3","Type":"ContainerStarted","Data":"4207e4585a57dba47b68bde432a5423db8472ec34525da41f794826e65a471f9"} Oct 13 09:08:06 crc kubenswrapper[4685]: I1013 09:08:06.251738 4685 generic.go:334] "Generic (PLEG): container finished" podID="9b1f2f60-a981-432b-a7cb-f84d1c4a80c3" containerID="4207e4585a57dba47b68bde432a5423db8472ec34525da41f794826e65a471f9" exitCode=0 Oct 13 09:08:06 crc kubenswrapper[4685]: I1013 09:08:06.251795 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qfp7" event={"ID":"9b1f2f60-a981-432b-a7cb-f84d1c4a80c3","Type":"ContainerDied","Data":"4207e4585a57dba47b68bde432a5423db8472ec34525da41f794826e65a471f9"} Oct 13 09:08:07 crc kubenswrapper[4685]: I1013 09:08:07.264382 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qfp7" event={"ID":"9b1f2f60-a981-432b-a7cb-f84d1c4a80c3","Type":"ContainerStarted","Data":"b168ebccdd722da72fb10d5740e21c8215ed334267cac53451fb7620e62a3489"} Oct 13 09:08:07 crc kubenswrapper[4685]: I1013 09:08:07.301842 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8qfp7" podStartSLOduration=1.6989538039999998 podStartE2EDuration="6.30182257s" podCreationTimestamp="2025-10-13 09:08:01 +0000 UTC" firstStartedPulling="2025-10-13 09:08:02.215230258 +0000 UTC m=+1407.363106019" lastFinishedPulling="2025-10-13 09:08:06.818099024 +0000 UTC m=+1411.965974785" observedRunningTime="2025-10-13 09:08:07.286338418 +0000 UTC m=+1412.434214179" watchObservedRunningTime="2025-10-13 09:08:07.30182257 +0000 UTC m=+1412.449698331" Oct 13 09:08:11 crc kubenswrapper[4685]: I1013 09:08:11.369441 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8qfp7" Oct 13 09:08:11 crc kubenswrapper[4685]: I1013 09:08:11.369927 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8qfp7" Oct 13 09:08:11 crc kubenswrapper[4685]: I1013 09:08:11.460310 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8qfp7" Oct 13 09:08:12 crc kubenswrapper[4685]: I1013 09:08:12.362418 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8qfp7" Oct 13 09:08:12 crc kubenswrapper[4685]: I1013 09:08:12.425726 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8qfp7"] Oct 13 09:08:12 crc kubenswrapper[4685]: I1013 09:08:12.432401 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xl64r" Oct 13 09:08:12 crc kubenswrapper[4685]: I1013 09:08:12.482433 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xl64r" Oct 13 09:08:14 crc kubenswrapper[4685]: I1013 09:08:14.097796 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xl64r"] Oct 13 09:08:14 crc kubenswrapper[4685]: I1013 09:08:14.329051 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xl64r" podUID="d4a4fa29-ba61-4eb0-94b1-d4d30daa135a" containerName="registry-server" containerID="cri-o://c84b91750d1f1ce6210d0fcb25296eac5754f23401474dfebd500c1101716adb" gracePeriod=2 Oct 13 09:08:14 crc kubenswrapper[4685]: I1013 09:08:14.329127 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8qfp7" podUID="9b1f2f60-a981-432b-a7cb-f84d1c4a80c3" containerName="registry-server" containerID="cri-o://b168ebccdd722da72fb10d5740e21c8215ed334267cac53451fb7620e62a3489" gracePeriod=2 Oct 13 09:08:14 crc kubenswrapper[4685]: I1013 09:08:14.839486 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xl64r" Oct 13 09:08:14 crc kubenswrapper[4685]: I1013 09:08:14.847170 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8qfp7" Oct 13 09:08:14 crc kubenswrapper[4685]: I1013 09:08:14.895990 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rd62w\" (UniqueName: \"kubernetes.io/projected/9b1f2f60-a981-432b-a7cb-f84d1c4a80c3-kube-api-access-rd62w\") pod \"9b1f2f60-a981-432b-a7cb-f84d1c4a80c3\" (UID: \"9b1f2f60-a981-432b-a7cb-f84d1c4a80c3\") " Oct 13 09:08:14 crc kubenswrapper[4685]: I1013 09:08:14.896173 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hn5f\" (UniqueName: \"kubernetes.io/projected/d4a4fa29-ba61-4eb0-94b1-d4d30daa135a-kube-api-access-7hn5f\") pod \"d4a4fa29-ba61-4eb0-94b1-d4d30daa135a\" (UID: \"d4a4fa29-ba61-4eb0-94b1-d4d30daa135a\") " Oct 13 09:08:14 crc kubenswrapper[4685]: I1013 09:08:14.896215 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b1f2f60-a981-432b-a7cb-f84d1c4a80c3-utilities\") pod \"9b1f2f60-a981-432b-a7cb-f84d1c4a80c3\" (UID: \"9b1f2f60-a981-432b-a7cb-f84d1c4a80c3\") " Oct 13 09:08:14 crc kubenswrapper[4685]: I1013 09:08:14.896234 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b1f2f60-a981-432b-a7cb-f84d1c4a80c3-catalog-content\") pod \"9b1f2f60-a981-432b-a7cb-f84d1c4a80c3\" (UID: \"9b1f2f60-a981-432b-a7cb-f84d1c4a80c3\") " Oct 13 09:08:14 crc kubenswrapper[4685]: I1013 09:08:14.896259 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a4fa29-ba61-4eb0-94b1-d4d30daa135a-catalog-content\") pod \"d4a4fa29-ba61-4eb0-94b1-d4d30daa135a\" (UID: \"d4a4fa29-ba61-4eb0-94b1-d4d30daa135a\") " Oct 13 09:08:14 crc kubenswrapper[4685]: I1013 09:08:14.896604 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a4fa29-ba61-4eb0-94b1-d4d30daa135a-utilities\") pod \"d4a4fa29-ba61-4eb0-94b1-d4d30daa135a\" (UID: \"d4a4fa29-ba61-4eb0-94b1-d4d30daa135a\") " Oct 13 09:08:14 crc kubenswrapper[4685]: I1013 09:08:14.898937 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4a4fa29-ba61-4eb0-94b1-d4d30daa135a-utilities" (OuterVolumeSpecName: "utilities") pod "d4a4fa29-ba61-4eb0-94b1-d4d30daa135a" (UID: "d4a4fa29-ba61-4eb0-94b1-d4d30daa135a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:08:14 crc kubenswrapper[4685]: I1013 09:08:14.899249 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b1f2f60-a981-432b-a7cb-f84d1c4a80c3-utilities" (OuterVolumeSpecName: "utilities") pod "9b1f2f60-a981-432b-a7cb-f84d1c4a80c3" (UID: "9b1f2f60-a981-432b-a7cb-f84d1c4a80c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:08:14 crc kubenswrapper[4685]: I1013 09:08:14.919251 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b1f2f60-a981-432b-a7cb-f84d1c4a80c3-kube-api-access-rd62w" (OuterVolumeSpecName: "kube-api-access-rd62w") pod "9b1f2f60-a981-432b-a7cb-f84d1c4a80c3" (UID: "9b1f2f60-a981-432b-a7cb-f84d1c4a80c3"). InnerVolumeSpecName "kube-api-access-rd62w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:08:14 crc kubenswrapper[4685]: I1013 09:08:14.927289 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4a4fa29-ba61-4eb0-94b1-d4d30daa135a-kube-api-access-7hn5f" (OuterVolumeSpecName: "kube-api-access-7hn5f") pod "d4a4fa29-ba61-4eb0-94b1-d4d30daa135a" (UID: "d4a4fa29-ba61-4eb0-94b1-d4d30daa135a"). InnerVolumeSpecName "kube-api-access-7hn5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:08:14 crc kubenswrapper[4685]: I1013 09:08:14.936429 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hn5f\" (UniqueName: \"kubernetes.io/projected/d4a4fa29-ba61-4eb0-94b1-d4d30daa135a-kube-api-access-7hn5f\") on node \"crc\" DevicePath \"\"" Oct 13 09:08:14 crc kubenswrapper[4685]: I1013 09:08:14.936455 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b1f2f60-a981-432b-a7cb-f84d1c4a80c3-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 09:08:14 crc kubenswrapper[4685]: I1013 09:08:14.936465 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a4fa29-ba61-4eb0-94b1-d4d30daa135a-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 09:08:14 crc kubenswrapper[4685]: I1013 09:08:14.936473 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rd62w\" (UniqueName: \"kubernetes.io/projected/9b1f2f60-a981-432b-a7cb-f84d1c4a80c3-kube-api-access-rd62w\") on node \"crc\" DevicePath \"\"" Oct 13 09:08:14 crc kubenswrapper[4685]: I1013 09:08:14.973954 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4a4fa29-ba61-4eb0-94b1-d4d30daa135a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4a4fa29-ba61-4eb0-94b1-d4d30daa135a" (UID: "d4a4fa29-ba61-4eb0-94b1-d4d30daa135a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:08:14 crc kubenswrapper[4685]: I1013 09:08:14.987592 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b1f2f60-a981-432b-a7cb-f84d1c4a80c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9b1f2f60-a981-432b-a7cb-f84d1c4a80c3" (UID: "9b1f2f60-a981-432b-a7cb-f84d1c4a80c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.038495 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b1f2f60-a981-432b-a7cb-f84d1c4a80c3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.038533 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a4fa29-ba61-4eb0-94b1-d4d30daa135a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.342201 4685 generic.go:334] "Generic (PLEG): container finished" podID="9b1f2f60-a981-432b-a7cb-f84d1c4a80c3" containerID="b168ebccdd722da72fb10d5740e21c8215ed334267cac53451fb7620e62a3489" exitCode=0 Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.342278 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8qfp7" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.342288 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qfp7" event={"ID":"9b1f2f60-a981-432b-a7cb-f84d1c4a80c3","Type":"ContainerDied","Data":"b168ebccdd722da72fb10d5740e21c8215ed334267cac53451fb7620e62a3489"} Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.343075 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qfp7" event={"ID":"9b1f2f60-a981-432b-a7cb-f84d1c4a80c3","Type":"ContainerDied","Data":"98d114306d773ede677073783ae9eff71840bd644145ef1b547cd8cfbb818148"} Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.343101 4685 scope.go:117] "RemoveContainer" containerID="b168ebccdd722da72fb10d5740e21c8215ed334267cac53451fb7620e62a3489" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.346427 4685 generic.go:334] "Generic (PLEG): container finished" podID="d4a4fa29-ba61-4eb0-94b1-d4d30daa135a" containerID="c84b91750d1f1ce6210d0fcb25296eac5754f23401474dfebd500c1101716adb" exitCode=0 Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.346473 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xl64r" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.346487 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xl64r" event={"ID":"d4a4fa29-ba61-4eb0-94b1-d4d30daa135a","Type":"ContainerDied","Data":"c84b91750d1f1ce6210d0fcb25296eac5754f23401474dfebd500c1101716adb"} Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.346804 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xl64r" event={"ID":"d4a4fa29-ba61-4eb0-94b1-d4d30daa135a","Type":"ContainerDied","Data":"30596a907873e5ea0e4a92c5623a40e6d978c655023dca684a01f684bf7ccb93"} Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.371792 4685 scope.go:117] "RemoveContainer" containerID="4207e4585a57dba47b68bde432a5423db8472ec34525da41f794826e65a471f9" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.401522 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8qfp7"] Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.414005 4685 scope.go:117] "RemoveContainer" containerID="46718462dbf280e9b5afb8268be15a0166ae6e020be5991ccd55ef8f449cf9c6" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.415806 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8qfp7"] Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.426735 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xl64r"] Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.434781 4685 scope.go:117] "RemoveContainer" containerID="b168ebccdd722da72fb10d5740e21c8215ed334267cac53451fb7620e62a3489" Oct 13 09:08:15 crc kubenswrapper[4685]: E1013 09:08:15.435292 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b168ebccdd722da72fb10d5740e21c8215ed334267cac53451fb7620e62a3489\": container with ID starting with b168ebccdd722da72fb10d5740e21c8215ed334267cac53451fb7620e62a3489 not found: ID does not exist" containerID="b168ebccdd722da72fb10d5740e21c8215ed334267cac53451fb7620e62a3489" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.435385 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b168ebccdd722da72fb10d5740e21c8215ed334267cac53451fb7620e62a3489"} err="failed to get container status \"b168ebccdd722da72fb10d5740e21c8215ed334267cac53451fb7620e62a3489\": rpc error: code = NotFound desc = could not find container \"b168ebccdd722da72fb10d5740e21c8215ed334267cac53451fb7620e62a3489\": container with ID starting with b168ebccdd722da72fb10d5740e21c8215ed334267cac53451fb7620e62a3489 not found: ID does not exist" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.435462 4685 scope.go:117] "RemoveContainer" containerID="4207e4585a57dba47b68bde432a5423db8472ec34525da41f794826e65a471f9" Oct 13 09:08:15 crc kubenswrapper[4685]: E1013 09:08:15.435859 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4207e4585a57dba47b68bde432a5423db8472ec34525da41f794826e65a471f9\": container with ID starting with 4207e4585a57dba47b68bde432a5423db8472ec34525da41f794826e65a471f9 not found: ID does not exist" containerID="4207e4585a57dba47b68bde432a5423db8472ec34525da41f794826e65a471f9" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.435944 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xl64r"] Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.435956 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4207e4585a57dba47b68bde432a5423db8472ec34525da41f794826e65a471f9"} err="failed to get container status \"4207e4585a57dba47b68bde432a5423db8472ec34525da41f794826e65a471f9\": rpc error: code = NotFound desc = could not find container \"4207e4585a57dba47b68bde432a5423db8472ec34525da41f794826e65a471f9\": container with ID starting with 4207e4585a57dba47b68bde432a5423db8472ec34525da41f794826e65a471f9 not found: ID does not exist" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.436006 4685 scope.go:117] "RemoveContainer" containerID="46718462dbf280e9b5afb8268be15a0166ae6e020be5991ccd55ef8f449cf9c6" Oct 13 09:08:15 crc kubenswrapper[4685]: E1013 09:08:15.436339 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46718462dbf280e9b5afb8268be15a0166ae6e020be5991ccd55ef8f449cf9c6\": container with ID starting with 46718462dbf280e9b5afb8268be15a0166ae6e020be5991ccd55ef8f449cf9c6 not found: ID does not exist" containerID="46718462dbf280e9b5afb8268be15a0166ae6e020be5991ccd55ef8f449cf9c6" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.436441 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46718462dbf280e9b5afb8268be15a0166ae6e020be5991ccd55ef8f449cf9c6"} err="failed to get container status \"46718462dbf280e9b5afb8268be15a0166ae6e020be5991ccd55ef8f449cf9c6\": rpc error: code = NotFound desc = could not find container \"46718462dbf280e9b5afb8268be15a0166ae6e020be5991ccd55ef8f449cf9c6\": container with ID starting with 46718462dbf280e9b5afb8268be15a0166ae6e020be5991ccd55ef8f449cf9c6 not found: ID does not exist" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.436503 4685 scope.go:117] "RemoveContainer" containerID="c84b91750d1f1ce6210d0fcb25296eac5754f23401474dfebd500c1101716adb" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.457888 4685 scope.go:117] "RemoveContainer" containerID="afb073b510b0d675317e4358488138193f1aeb7708e1f3b8b34069dd785cdddf" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.483200 4685 scope.go:117] "RemoveContainer" containerID="e8ccecbf9cae068f16592f61323b69de10a9f7f1116a911467be425f35bedd44" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.534937 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b1f2f60-a981-432b-a7cb-f84d1c4a80c3" path="/var/lib/kubelet/pods/9b1f2f60-a981-432b-a7cb-f84d1c4a80c3/volumes" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.535770 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4a4fa29-ba61-4eb0-94b1-d4d30daa135a" path="/var/lib/kubelet/pods/d4a4fa29-ba61-4eb0-94b1-d4d30daa135a/volumes" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.572822 4685 scope.go:117] "RemoveContainer" containerID="c84b91750d1f1ce6210d0fcb25296eac5754f23401474dfebd500c1101716adb" Oct 13 09:08:15 crc kubenswrapper[4685]: E1013 09:08:15.580478 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c84b91750d1f1ce6210d0fcb25296eac5754f23401474dfebd500c1101716adb\": container with ID starting with c84b91750d1f1ce6210d0fcb25296eac5754f23401474dfebd500c1101716adb not found: ID does not exist" containerID="c84b91750d1f1ce6210d0fcb25296eac5754f23401474dfebd500c1101716adb" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.580516 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c84b91750d1f1ce6210d0fcb25296eac5754f23401474dfebd500c1101716adb"} err="failed to get container status \"c84b91750d1f1ce6210d0fcb25296eac5754f23401474dfebd500c1101716adb\": rpc error: code = NotFound desc = could not find container \"c84b91750d1f1ce6210d0fcb25296eac5754f23401474dfebd500c1101716adb\": container with ID starting with c84b91750d1f1ce6210d0fcb25296eac5754f23401474dfebd500c1101716adb not found: ID does not exist" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.580544 4685 scope.go:117] "RemoveContainer" containerID="afb073b510b0d675317e4358488138193f1aeb7708e1f3b8b34069dd785cdddf" Oct 13 09:08:15 crc kubenswrapper[4685]: E1013 09:08:15.588700 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afb073b510b0d675317e4358488138193f1aeb7708e1f3b8b34069dd785cdddf\": container with ID starting with afb073b510b0d675317e4358488138193f1aeb7708e1f3b8b34069dd785cdddf not found: ID does not exist" containerID="afb073b510b0d675317e4358488138193f1aeb7708e1f3b8b34069dd785cdddf" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.588749 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afb073b510b0d675317e4358488138193f1aeb7708e1f3b8b34069dd785cdddf"} err="failed to get container status \"afb073b510b0d675317e4358488138193f1aeb7708e1f3b8b34069dd785cdddf\": rpc error: code = NotFound desc = could not find container \"afb073b510b0d675317e4358488138193f1aeb7708e1f3b8b34069dd785cdddf\": container with ID starting with afb073b510b0d675317e4358488138193f1aeb7708e1f3b8b34069dd785cdddf not found: ID does not exist" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.588784 4685 scope.go:117] "RemoveContainer" containerID="e8ccecbf9cae068f16592f61323b69de10a9f7f1116a911467be425f35bedd44" Oct 13 09:08:15 crc kubenswrapper[4685]: E1013 09:08:15.593026 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8ccecbf9cae068f16592f61323b69de10a9f7f1116a911467be425f35bedd44\": container with ID starting with e8ccecbf9cae068f16592f61323b69de10a9f7f1116a911467be425f35bedd44 not found: ID does not exist" containerID="e8ccecbf9cae068f16592f61323b69de10a9f7f1116a911467be425f35bedd44" Oct 13 09:08:15 crc kubenswrapper[4685]: I1013 09:08:15.593066 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8ccecbf9cae068f16592f61323b69de10a9f7f1116a911467be425f35bedd44"} err="failed to get container status \"e8ccecbf9cae068f16592f61323b69de10a9f7f1116a911467be425f35bedd44\": rpc error: code = NotFound desc = could not find container \"e8ccecbf9cae068f16592f61323b69de10a9f7f1116a911467be425f35bedd44\": container with ID starting with e8ccecbf9cae068f16592f61323b69de10a9f7f1116a911467be425f35bedd44 not found: ID does not exist" Oct 13 09:08:22 crc kubenswrapper[4685]: I1013 09:08:22.980361 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:08:22 crc kubenswrapper[4685]: I1013 09:08:22.982430 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:08:25 crc kubenswrapper[4685]: I1013 09:08:25.657042 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7vrsq"] Oct 13 09:08:25 crc kubenswrapper[4685]: E1013 09:08:25.657820 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4a4fa29-ba61-4eb0-94b1-d4d30daa135a" containerName="registry-server" Oct 13 09:08:25 crc kubenswrapper[4685]: I1013 09:08:25.657835 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a4fa29-ba61-4eb0-94b1-d4d30daa135a" containerName="registry-server" Oct 13 09:08:25 crc kubenswrapper[4685]: E1013 09:08:25.657863 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4a4fa29-ba61-4eb0-94b1-d4d30daa135a" containerName="extract-utilities" Oct 13 09:08:25 crc kubenswrapper[4685]: I1013 09:08:25.657871 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a4fa29-ba61-4eb0-94b1-d4d30daa135a" containerName="extract-utilities" Oct 13 09:08:25 crc kubenswrapper[4685]: E1013 09:08:25.657895 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b1f2f60-a981-432b-a7cb-f84d1c4a80c3" containerName="extract-utilities" Oct 13 09:08:25 crc kubenswrapper[4685]: I1013 09:08:25.657903 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b1f2f60-a981-432b-a7cb-f84d1c4a80c3" containerName="extract-utilities" Oct 13 09:08:25 crc kubenswrapper[4685]: E1013 09:08:25.657944 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b1f2f60-a981-432b-a7cb-f84d1c4a80c3" containerName="extract-content" Oct 13 09:08:25 crc kubenswrapper[4685]: I1013 09:08:25.657952 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b1f2f60-a981-432b-a7cb-f84d1c4a80c3" containerName="extract-content" Oct 13 09:08:25 crc kubenswrapper[4685]: E1013 09:08:25.657966 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b1f2f60-a981-432b-a7cb-f84d1c4a80c3" containerName="registry-server" Oct 13 09:08:25 crc kubenswrapper[4685]: I1013 09:08:25.657974 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b1f2f60-a981-432b-a7cb-f84d1c4a80c3" containerName="registry-server" Oct 13 09:08:25 crc kubenswrapper[4685]: E1013 09:08:25.657996 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4a4fa29-ba61-4eb0-94b1-d4d30daa135a" containerName="extract-content" Oct 13 09:08:25 crc kubenswrapper[4685]: I1013 09:08:25.658003 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a4fa29-ba61-4eb0-94b1-d4d30daa135a" containerName="extract-content" Oct 13 09:08:25 crc kubenswrapper[4685]: I1013 09:08:25.658207 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4a4fa29-ba61-4eb0-94b1-d4d30daa135a" containerName="registry-server" Oct 13 09:08:25 crc kubenswrapper[4685]: I1013 09:08:25.658223 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b1f2f60-a981-432b-a7cb-f84d1c4a80c3" containerName="registry-server" Oct 13 09:08:25 crc kubenswrapper[4685]: I1013 09:08:25.659885 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7vrsq" Oct 13 09:08:25 crc kubenswrapper[4685]: I1013 09:08:25.676289 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7vrsq"] Oct 13 09:08:25 crc kubenswrapper[4685]: I1013 09:08:25.751292 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1-utilities\") pod \"redhat-operators-7vrsq\" (UID: \"1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1\") " pod="openshift-marketplace/redhat-operators-7vrsq" Oct 13 09:08:25 crc kubenswrapper[4685]: I1013 09:08:25.751374 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1-catalog-content\") pod \"redhat-operators-7vrsq\" (UID: \"1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1\") " pod="openshift-marketplace/redhat-operators-7vrsq" Oct 13 09:08:25 crc kubenswrapper[4685]: I1013 09:08:25.751644 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwpl7\" (UniqueName: \"kubernetes.io/projected/1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1-kube-api-access-mwpl7\") pod \"redhat-operators-7vrsq\" (UID: \"1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1\") " pod="openshift-marketplace/redhat-operators-7vrsq" Oct 13 09:08:25 crc kubenswrapper[4685]: I1013 09:08:25.853718 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1-catalog-content\") pod \"redhat-operators-7vrsq\" (UID: \"1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1\") " pod="openshift-marketplace/redhat-operators-7vrsq" Oct 13 09:08:25 crc kubenswrapper[4685]: I1013 09:08:25.853844 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwpl7\" (UniqueName: \"kubernetes.io/projected/1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1-kube-api-access-mwpl7\") pod \"redhat-operators-7vrsq\" (UID: \"1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1\") " pod="openshift-marketplace/redhat-operators-7vrsq" Oct 13 09:08:25 crc kubenswrapper[4685]: I1013 09:08:25.853930 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1-utilities\") pod \"redhat-operators-7vrsq\" (UID: \"1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1\") " pod="openshift-marketplace/redhat-operators-7vrsq" Oct 13 09:08:25 crc kubenswrapper[4685]: I1013 09:08:25.854345 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1-utilities\") pod \"redhat-operators-7vrsq\" (UID: \"1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1\") " pod="openshift-marketplace/redhat-operators-7vrsq" Oct 13 09:08:25 crc kubenswrapper[4685]: I1013 09:08:25.854567 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1-catalog-content\") pod \"redhat-operators-7vrsq\" (UID: \"1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1\") " pod="openshift-marketplace/redhat-operators-7vrsq" Oct 13 09:08:25 crc kubenswrapper[4685]: I1013 09:08:25.886656 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwpl7\" (UniqueName: \"kubernetes.io/projected/1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1-kube-api-access-mwpl7\") pod \"redhat-operators-7vrsq\" (UID: \"1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1\") " pod="openshift-marketplace/redhat-operators-7vrsq" Oct 13 09:08:25 crc kubenswrapper[4685]: I1013 09:08:25.981409 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7vrsq" Oct 13 09:08:26 crc kubenswrapper[4685]: I1013 09:08:26.432935 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7vrsq"] Oct 13 09:08:26 crc kubenswrapper[4685]: I1013 09:08:26.472535 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vrsq" event={"ID":"1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1","Type":"ContainerStarted","Data":"397b427f43ac533f1095f6b309e3f89d66c8b6408d0a451b4963c61919e95d5c"} Oct 13 09:08:27 crc kubenswrapper[4685]: I1013 09:08:27.486772 4685 generic.go:334] "Generic (PLEG): container finished" podID="1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1" containerID="ca081f4e329ac14663289b842602fd56f0f7c48f3a8830316b38e6126ad6eb29" exitCode=0 Oct 13 09:08:27 crc kubenswrapper[4685]: I1013 09:08:27.487095 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vrsq" event={"ID":"1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1","Type":"ContainerDied","Data":"ca081f4e329ac14663289b842602fd56f0f7c48f3a8830316b38e6126ad6eb29"} Oct 13 09:08:29 crc kubenswrapper[4685]: I1013 09:08:29.521139 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vrsq" event={"ID":"1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1","Type":"ContainerStarted","Data":"cce1d6503cf69cf777c8e70c02163e93e0a4d0bee079d4c4329fd6257b4b9abe"} Oct 13 09:08:37 crc kubenswrapper[4685]: I1013 09:08:37.609999 4685 generic.go:334] "Generic (PLEG): container finished" podID="1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1" containerID="cce1d6503cf69cf777c8e70c02163e93e0a4d0bee079d4c4329fd6257b4b9abe" exitCode=0 Oct 13 09:08:37 crc kubenswrapper[4685]: I1013 09:08:37.610037 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vrsq" event={"ID":"1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1","Type":"ContainerDied","Data":"cce1d6503cf69cf777c8e70c02163e93e0a4d0bee079d4c4329fd6257b4b9abe"} Oct 13 09:08:38 crc kubenswrapper[4685]: I1013 09:08:38.632324 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vrsq" event={"ID":"1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1","Type":"ContainerStarted","Data":"e07d492ae5f76cae59469663eb740bef30a73742ab3e6dc87e46b570863123a1"} Oct 13 09:08:38 crc kubenswrapper[4685]: I1013 09:08:38.659137 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7vrsq" podStartSLOduration=3.122533946 podStartE2EDuration="13.659118519s" podCreationTimestamp="2025-10-13 09:08:25 +0000 UTC" firstStartedPulling="2025-10-13 09:08:27.489027614 +0000 UTC m=+1432.636903375" lastFinishedPulling="2025-10-13 09:08:38.025612187 +0000 UTC m=+1443.173487948" observedRunningTime="2025-10-13 09:08:38.65070342 +0000 UTC m=+1443.798579201" watchObservedRunningTime="2025-10-13 09:08:38.659118519 +0000 UTC m=+1443.806994290" Oct 13 09:08:45 crc kubenswrapper[4685]: I1013 09:08:45.982456 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7vrsq" Oct 13 09:08:45 crc kubenswrapper[4685]: I1013 09:08:45.983010 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7vrsq" Oct 13 09:08:47 crc kubenswrapper[4685]: I1013 09:08:47.025581 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7vrsq" podUID="1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1" containerName="registry-server" probeResult="failure" output=< Oct 13 09:08:47 crc kubenswrapper[4685]: timeout: failed to connect service ":50051" within 1s Oct 13 09:08:47 crc kubenswrapper[4685]: > Oct 13 09:08:52 crc kubenswrapper[4685]: I1013 09:08:52.979971 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:08:52 crc kubenswrapper[4685]: I1013 09:08:52.980518 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:08:54 crc kubenswrapper[4685]: I1013 09:08:54.187768 4685 scope.go:117] "RemoveContainer" containerID="2ea1dee834b2cf10e4663b9d31c46ab5dadb18dcb165e4a28a827fffab9359fe" Oct 13 09:08:54 crc kubenswrapper[4685]: I1013 09:08:54.210242 4685 scope.go:117] "RemoveContainer" containerID="c4e5d25acb56a14f5e7c988629b99798caf9ed0571625e21721d9f721e1e717a" Oct 13 09:08:56 crc kubenswrapper[4685]: I1013 09:08:56.031666 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7vrsq" Oct 13 09:08:56 crc kubenswrapper[4685]: I1013 09:08:56.093086 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7vrsq" Oct 13 09:08:56 crc kubenswrapper[4685]: I1013 09:08:56.860780 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7vrsq"] Oct 13 09:08:57 crc kubenswrapper[4685]: I1013 09:08:57.795902 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7vrsq" podUID="1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1" containerName="registry-server" containerID="cri-o://e07d492ae5f76cae59469663eb740bef30a73742ab3e6dc87e46b570863123a1" gracePeriod=2 Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.337042 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7vrsq" Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.453348 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwpl7\" (UniqueName: \"kubernetes.io/projected/1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1-kube-api-access-mwpl7\") pod \"1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1\" (UID: \"1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1\") " Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.453499 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1-catalog-content\") pod \"1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1\" (UID: \"1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1\") " Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.453664 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1-utilities\") pod \"1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1\" (UID: \"1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1\") " Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.454290 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1-utilities" (OuterVolumeSpecName: "utilities") pod "1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1" (UID: "1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.454738 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.462437 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1-kube-api-access-mwpl7" (OuterVolumeSpecName: "kube-api-access-mwpl7") pod "1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1" (UID: "1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1"). InnerVolumeSpecName "kube-api-access-mwpl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.531880 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1" (UID: "1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.556773 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwpl7\" (UniqueName: \"kubernetes.io/projected/1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1-kube-api-access-mwpl7\") on node \"crc\" DevicePath \"\"" Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.556810 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.806245 4685 generic.go:334] "Generic (PLEG): container finished" podID="1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1" containerID="e07d492ae5f76cae59469663eb740bef30a73742ab3e6dc87e46b570863123a1" exitCode=0 Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.806284 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vrsq" event={"ID":"1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1","Type":"ContainerDied","Data":"e07d492ae5f76cae59469663eb740bef30a73742ab3e6dc87e46b570863123a1"} Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.806306 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vrsq" event={"ID":"1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1","Type":"ContainerDied","Data":"397b427f43ac533f1095f6b309e3f89d66c8b6408d0a451b4963c61919e95d5c"} Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.806313 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7vrsq" Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.806320 4685 scope.go:117] "RemoveContainer" containerID="e07d492ae5f76cae59469663eb740bef30a73742ab3e6dc87e46b570863123a1" Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.823743 4685 scope.go:117] "RemoveContainer" containerID="cce1d6503cf69cf777c8e70c02163e93e0a4d0bee079d4c4329fd6257b4b9abe" Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.842733 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7vrsq"] Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.852819 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7vrsq"] Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.855110 4685 scope.go:117] "RemoveContainer" containerID="ca081f4e329ac14663289b842602fd56f0f7c48f3a8830316b38e6126ad6eb29" Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.901574 4685 scope.go:117] "RemoveContainer" containerID="e07d492ae5f76cae59469663eb740bef30a73742ab3e6dc87e46b570863123a1" Oct 13 09:08:58 crc kubenswrapper[4685]: E1013 09:08:58.902130 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e07d492ae5f76cae59469663eb740bef30a73742ab3e6dc87e46b570863123a1\": container with ID starting with e07d492ae5f76cae59469663eb740bef30a73742ab3e6dc87e46b570863123a1 not found: ID does not exist" containerID="e07d492ae5f76cae59469663eb740bef30a73742ab3e6dc87e46b570863123a1" Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.902170 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e07d492ae5f76cae59469663eb740bef30a73742ab3e6dc87e46b570863123a1"} err="failed to get container status \"e07d492ae5f76cae59469663eb740bef30a73742ab3e6dc87e46b570863123a1\": rpc error: code = NotFound desc = could not find container \"e07d492ae5f76cae59469663eb740bef30a73742ab3e6dc87e46b570863123a1\": container with ID starting with e07d492ae5f76cae59469663eb740bef30a73742ab3e6dc87e46b570863123a1 not found: ID does not exist" Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.902196 4685 scope.go:117] "RemoveContainer" containerID="cce1d6503cf69cf777c8e70c02163e93e0a4d0bee079d4c4329fd6257b4b9abe" Oct 13 09:08:58 crc kubenswrapper[4685]: E1013 09:08:58.902591 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cce1d6503cf69cf777c8e70c02163e93e0a4d0bee079d4c4329fd6257b4b9abe\": container with ID starting with cce1d6503cf69cf777c8e70c02163e93e0a4d0bee079d4c4329fd6257b4b9abe not found: ID does not exist" containerID="cce1d6503cf69cf777c8e70c02163e93e0a4d0bee079d4c4329fd6257b4b9abe" Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.902624 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cce1d6503cf69cf777c8e70c02163e93e0a4d0bee079d4c4329fd6257b4b9abe"} err="failed to get container status \"cce1d6503cf69cf777c8e70c02163e93e0a4d0bee079d4c4329fd6257b4b9abe\": rpc error: code = NotFound desc = could not find container \"cce1d6503cf69cf777c8e70c02163e93e0a4d0bee079d4c4329fd6257b4b9abe\": container with ID starting with cce1d6503cf69cf777c8e70c02163e93e0a4d0bee079d4c4329fd6257b4b9abe not found: ID does not exist" Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.902644 4685 scope.go:117] "RemoveContainer" containerID="ca081f4e329ac14663289b842602fd56f0f7c48f3a8830316b38e6126ad6eb29" Oct 13 09:08:58 crc kubenswrapper[4685]: E1013 09:08:58.902929 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca081f4e329ac14663289b842602fd56f0f7c48f3a8830316b38e6126ad6eb29\": container with ID starting with ca081f4e329ac14663289b842602fd56f0f7c48f3a8830316b38e6126ad6eb29 not found: ID does not exist" containerID="ca081f4e329ac14663289b842602fd56f0f7c48f3a8830316b38e6126ad6eb29" Oct 13 09:08:58 crc kubenswrapper[4685]: I1013 09:08:58.902980 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca081f4e329ac14663289b842602fd56f0f7c48f3a8830316b38e6126ad6eb29"} err="failed to get container status \"ca081f4e329ac14663289b842602fd56f0f7c48f3a8830316b38e6126ad6eb29\": rpc error: code = NotFound desc = could not find container \"ca081f4e329ac14663289b842602fd56f0f7c48f3a8830316b38e6126ad6eb29\": container with ID starting with ca081f4e329ac14663289b842602fd56f0f7c48f3a8830316b38e6126ad6eb29 not found: ID does not exist" Oct 13 09:08:59 crc kubenswrapper[4685]: I1013 09:08:59.523403 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1" path="/var/lib/kubelet/pods/1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1/volumes" Oct 13 09:09:22 crc kubenswrapper[4685]: I1013 09:09:22.980986 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:09:22 crc kubenswrapper[4685]: I1013 09:09:22.983805 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:09:22 crc kubenswrapper[4685]: I1013 09:09:22.983903 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 09:09:22 crc kubenswrapper[4685]: I1013 09:09:22.986530 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a2277b2f3c48d695283091e6c5552e7453d3c307f3bc3a11a7a1f02c1122c7c2"} pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 09:09:22 crc kubenswrapper[4685]: I1013 09:09:22.986818 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" containerID="cri-o://a2277b2f3c48d695283091e6c5552e7453d3c307f3bc3a11a7a1f02c1122c7c2" gracePeriod=600 Oct 13 09:09:24 crc kubenswrapper[4685]: I1013 09:09:24.060957 4685 generic.go:334] "Generic (PLEG): container finished" podID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerID="a2277b2f3c48d695283091e6c5552e7453d3c307f3bc3a11a7a1f02c1122c7c2" exitCode=0 Oct 13 09:09:24 crc kubenswrapper[4685]: I1013 09:09:24.061370 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerDied","Data":"a2277b2f3c48d695283091e6c5552e7453d3c307f3bc3a11a7a1f02c1122c7c2"} Oct 13 09:09:24 crc kubenswrapper[4685]: I1013 09:09:24.061412 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerStarted","Data":"199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570"} Oct 13 09:09:24 crc kubenswrapper[4685]: I1013 09:09:24.061440 4685 scope.go:117] "RemoveContainer" containerID="3648096a0845e8e530c524922c94cf62c1c3b816bc3674fd046d4900ce16a389" Oct 13 09:09:50 crc kubenswrapper[4685]: I1013 09:09:50.332682 4685 generic.go:334] "Generic (PLEG): container finished" podID="b001d17a-1aea-44ba-86c5-ba6b312156c1" containerID="cedf68f34af27ba6b0e7b6a3ebc8c368caa3b020eacec7a71ecf4f542d94655f" exitCode=1 Oct 13 09:09:50 crc kubenswrapper[4685]: I1013 09:09:50.332868 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" event={"ID":"b001d17a-1aea-44ba-86c5-ba6b312156c1","Type":"ContainerDied","Data":"cedf68f34af27ba6b0e7b6a3ebc8c368caa3b020eacec7a71ecf4f542d94655f"} Oct 13 09:09:50 crc kubenswrapper[4685]: I1013 09:09:50.333333 4685 scope.go:117] "RemoveContainer" containerID="6f8ac1889cdb78c30c1167f41924fc8421f395f14a8be5ba49377fbd32fe2472" Oct 13 09:09:50 crc kubenswrapper[4685]: I1013 09:09:50.334171 4685 scope.go:117] "RemoveContainer" containerID="cedf68f34af27ba6b0e7b6a3ebc8c368caa3b020eacec7a71ecf4f542d94655f" Oct 13 09:09:50 crc kubenswrapper[4685]: E1013 09:09:50.334599 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 40s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:09:53 crc kubenswrapper[4685]: I1013 09:09:53.443090 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:09:53 crc kubenswrapper[4685]: I1013 09:09:53.444666 4685 scope.go:117] "RemoveContainer" containerID="cedf68f34af27ba6b0e7b6a3ebc8c368caa3b020eacec7a71ecf4f542d94655f" Oct 13 09:09:53 crc kubenswrapper[4685]: E1013 09:09:53.445193 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 40s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:10:03 crc kubenswrapper[4685]: I1013 09:10:03.443124 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:10:03 crc kubenswrapper[4685]: I1013 09:10:03.446468 4685 scope.go:117] "RemoveContainer" containerID="cedf68f34af27ba6b0e7b6a3ebc8c368caa3b020eacec7a71ecf4f542d94655f" Oct 13 09:10:03 crc kubenswrapper[4685]: E1013 09:10:03.447180 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 40s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:10:16 crc kubenswrapper[4685]: I1013 09:10:16.502584 4685 scope.go:117] "RemoveContainer" containerID="cedf68f34af27ba6b0e7b6a3ebc8c368caa3b020eacec7a71ecf4f542d94655f" Oct 13 09:10:16 crc kubenswrapper[4685]: E1013 09:10:16.503307 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 40s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:10:27 crc kubenswrapper[4685]: I1013 09:10:27.503308 4685 scope.go:117] "RemoveContainer" containerID="cedf68f34af27ba6b0e7b6a3ebc8c368caa3b020eacec7a71ecf4f542d94655f" Oct 13 09:10:27 crc kubenswrapper[4685]: E1013 09:10:27.504139 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 40s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:10:38 crc kubenswrapper[4685]: I1013 09:10:38.503106 4685 scope.go:117] "RemoveContainer" containerID="cedf68f34af27ba6b0e7b6a3ebc8c368caa3b020eacec7a71ecf4f542d94655f" Oct 13 09:10:39 crc kubenswrapper[4685]: I1013 09:10:39.812720 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" event={"ID":"b001d17a-1aea-44ba-86c5-ba6b312156c1","Type":"ContainerStarted","Data":"f4bd2015aa86f0bf7c40324d231108b8eefc9288b86f37bc304379cab4fbde84"} Oct 13 09:10:39 crc kubenswrapper[4685]: I1013 09:10:39.813518 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:10:41 crc kubenswrapper[4685]: I1013 09:10:41.069056 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-jt65g"] Oct 13 09:10:41 crc kubenswrapper[4685]: I1013 09:10:41.081492 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-bccdb"] Oct 13 09:10:41 crc kubenswrapper[4685]: I1013 09:10:41.088776 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-jt65g"] Oct 13 09:10:41 crc kubenswrapper[4685]: I1013 09:10:41.096026 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-bccdb"] Oct 13 09:10:41 crc kubenswrapper[4685]: I1013 09:10:41.529834 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6334858c-1809-484a-a74f-49e28a18e9b2" path="/var/lib/kubelet/pods/6334858c-1809-484a-a74f-49e28a18e9b2/volumes" Oct 13 09:10:41 crc kubenswrapper[4685]: I1013 09:10:41.563668 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a48e241f-11b8-4c11-a17a-0d96b2b37df2" path="/var/lib/kubelet/pods/a48e241f-11b8-4c11-a17a-0d96b2b37df2/volumes" Oct 13 09:10:42 crc kubenswrapper[4685]: I1013 09:10:42.844175 4685 generic.go:334] "Generic (PLEG): container finished" podID="5056abd4-630c-49fe-9ddf-1294f0f6b55a" containerID="6131eec83299a39e69b777f0edd29636b60749847c97c68d2af25429890bdd58" exitCode=0 Oct 13 09:10:42 crc kubenswrapper[4685]: I1013 09:10:42.844225 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl" event={"ID":"5056abd4-630c-49fe-9ddf-1294f0f6b55a","Type":"ContainerDied","Data":"6131eec83299a39e69b777f0edd29636b60749847c97c68d2af25429890bdd58"} Oct 13 09:10:43 crc kubenswrapper[4685]: I1013 09:10:43.446385 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.441630 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.626817 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5056abd4-630c-49fe-9ddf-1294f0f6b55a-ssh-key\") pod \"5056abd4-630c-49fe-9ddf-1294f0f6b55a\" (UID: \"5056abd4-630c-49fe-9ddf-1294f0f6b55a\") " Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.626885 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxgn8\" (UniqueName: \"kubernetes.io/projected/5056abd4-630c-49fe-9ddf-1294f0f6b55a-kube-api-access-qxgn8\") pod \"5056abd4-630c-49fe-9ddf-1294f0f6b55a\" (UID: \"5056abd4-630c-49fe-9ddf-1294f0f6b55a\") " Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.627041 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5056abd4-630c-49fe-9ddf-1294f0f6b55a-bootstrap-combined-ca-bundle\") pod \"5056abd4-630c-49fe-9ddf-1294f0f6b55a\" (UID: \"5056abd4-630c-49fe-9ddf-1294f0f6b55a\") " Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.627067 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5056abd4-630c-49fe-9ddf-1294f0f6b55a-inventory\") pod \"5056abd4-630c-49fe-9ddf-1294f0f6b55a\" (UID: \"5056abd4-630c-49fe-9ddf-1294f0f6b55a\") " Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.633434 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5056abd4-630c-49fe-9ddf-1294f0f6b55a-kube-api-access-qxgn8" (OuterVolumeSpecName: "kube-api-access-qxgn8") pod "5056abd4-630c-49fe-9ddf-1294f0f6b55a" (UID: "5056abd4-630c-49fe-9ddf-1294f0f6b55a"). InnerVolumeSpecName "kube-api-access-qxgn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.635195 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5056abd4-630c-49fe-9ddf-1294f0f6b55a-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "5056abd4-630c-49fe-9ddf-1294f0f6b55a" (UID: "5056abd4-630c-49fe-9ddf-1294f0f6b55a"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.658839 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5056abd4-630c-49fe-9ddf-1294f0f6b55a-inventory" (OuterVolumeSpecName: "inventory") pod "5056abd4-630c-49fe-9ddf-1294f0f6b55a" (UID: "5056abd4-630c-49fe-9ddf-1294f0f6b55a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.660478 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5056abd4-630c-49fe-9ddf-1294f0f6b55a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5056abd4-630c-49fe-9ddf-1294f0f6b55a" (UID: "5056abd4-630c-49fe-9ddf-1294f0f6b55a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.729685 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5056abd4-630c-49fe-9ddf-1294f0f6b55a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.729730 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxgn8\" (UniqueName: \"kubernetes.io/projected/5056abd4-630c-49fe-9ddf-1294f0f6b55a-kube-api-access-qxgn8\") on node \"crc\" DevicePath \"\"" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.729744 4685 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5056abd4-630c-49fe-9ddf-1294f0f6b55a-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.729753 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5056abd4-630c-49fe-9ddf-1294f0f6b55a-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.867855 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.867859 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl" event={"ID":"5056abd4-630c-49fe-9ddf-1294f0f6b55a","Type":"ContainerDied","Data":"9ae9f3547a9800eccb87ab49c090b7f0ccecaac0ffa1100ef286bfe7569b52ff"} Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.867958 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ae9f3547a9800eccb87ab49c090b7f0ccecaac0ffa1100ef286bfe7569b52ff" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.955973 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc"] Oct 13 09:10:44 crc kubenswrapper[4685]: E1013 09:10:44.956415 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5056abd4-630c-49fe-9ddf-1294f0f6b55a" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.956431 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="5056abd4-630c-49fe-9ddf-1294f0f6b55a" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 13 09:10:44 crc kubenswrapper[4685]: E1013 09:10:44.956448 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1" containerName="extract-utilities" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.956455 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1" containerName="extract-utilities" Oct 13 09:10:44 crc kubenswrapper[4685]: E1013 09:10:44.956466 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1" containerName="registry-server" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.956471 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1" containerName="registry-server" Oct 13 09:10:44 crc kubenswrapper[4685]: E1013 09:10:44.956483 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1" containerName="extract-content" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.956489 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1" containerName="extract-content" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.956669 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cc9ff89-a4f7-4a41-8bfd-204a368aa7d1" containerName="registry-server" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.956683 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="5056abd4-630c-49fe-9ddf-1294f0f6b55a" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.957309 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.959735 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.960065 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.960245 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hcn4z" Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.964928 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc"] Oct 13 09:10:44 crc kubenswrapper[4685]: I1013 09:10:44.968286 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 09:10:45 crc kubenswrapper[4685]: I1013 09:10:45.037328 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmfnx\" (UniqueName: \"kubernetes.io/projected/49edf7b8-3834-40c0-a038-e6c2388842da-kube-api-access-tmfnx\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc\" (UID: \"49edf7b8-3834-40c0-a038-e6c2388842da\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc" Oct 13 09:10:45 crc kubenswrapper[4685]: I1013 09:10:45.037392 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49edf7b8-3834-40c0-a038-e6c2388842da-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc\" (UID: \"49edf7b8-3834-40c0-a038-e6c2388842da\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc" Oct 13 09:10:45 crc kubenswrapper[4685]: I1013 09:10:45.037508 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/49edf7b8-3834-40c0-a038-e6c2388842da-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc\" (UID: \"49edf7b8-3834-40c0-a038-e6c2388842da\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc" Oct 13 09:10:45 crc kubenswrapper[4685]: I1013 09:10:45.140061 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmfnx\" (UniqueName: \"kubernetes.io/projected/49edf7b8-3834-40c0-a038-e6c2388842da-kube-api-access-tmfnx\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc\" (UID: \"49edf7b8-3834-40c0-a038-e6c2388842da\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc" Oct 13 09:10:45 crc kubenswrapper[4685]: I1013 09:10:45.140865 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49edf7b8-3834-40c0-a038-e6c2388842da-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc\" (UID: \"49edf7b8-3834-40c0-a038-e6c2388842da\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc" Oct 13 09:10:45 crc kubenswrapper[4685]: I1013 09:10:45.143465 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/49edf7b8-3834-40c0-a038-e6c2388842da-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc\" (UID: \"49edf7b8-3834-40c0-a038-e6c2388842da\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc" Oct 13 09:10:45 crc kubenswrapper[4685]: I1013 09:10:45.146964 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49edf7b8-3834-40c0-a038-e6c2388842da-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc\" (UID: \"49edf7b8-3834-40c0-a038-e6c2388842da\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc" Oct 13 09:10:45 crc kubenswrapper[4685]: I1013 09:10:45.146967 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/49edf7b8-3834-40c0-a038-e6c2388842da-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc\" (UID: \"49edf7b8-3834-40c0-a038-e6c2388842da\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc" Oct 13 09:10:45 crc kubenswrapper[4685]: I1013 09:10:45.163967 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmfnx\" (UniqueName: \"kubernetes.io/projected/49edf7b8-3834-40c0-a038-e6c2388842da-kube-api-access-tmfnx\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc\" (UID: \"49edf7b8-3834-40c0-a038-e6c2388842da\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc" Oct 13 09:10:45 crc kubenswrapper[4685]: I1013 09:10:45.279827 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc" Oct 13 09:10:45 crc kubenswrapper[4685]: I1013 09:10:45.636505 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc"] Oct 13 09:10:45 crc kubenswrapper[4685]: I1013 09:10:45.644081 4685 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 13 09:10:45 crc kubenswrapper[4685]: I1013 09:10:45.880899 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc" event={"ID":"49edf7b8-3834-40c0-a038-e6c2388842da","Type":"ContainerStarted","Data":"e70170c92131416a3c540af44018baa4a452f6b1d7d8d86f3672ad3bb8d01d5f"} Oct 13 09:10:46 crc kubenswrapper[4685]: I1013 09:10:46.893981 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc" event={"ID":"49edf7b8-3834-40c0-a038-e6c2388842da","Type":"ContainerStarted","Data":"cacdd8bf2dfff3338c8583d2dc7593c3179d46ad3f96585eae1e55a75337df58"} Oct 13 09:10:47 crc kubenswrapper[4685]: I1013 09:10:47.025170 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc" podStartSLOduration=2.493726316 podStartE2EDuration="3.025148701s" podCreationTimestamp="2025-10-13 09:10:44 +0000 UTC" firstStartedPulling="2025-10-13 09:10:45.643780114 +0000 UTC m=+1570.791655875" lastFinishedPulling="2025-10-13 09:10:46.175202489 +0000 UTC m=+1571.323078260" observedRunningTime="2025-10-13 09:10:46.912382682 +0000 UTC m=+1572.060258463" watchObservedRunningTime="2025-10-13 09:10:47.025148701 +0000 UTC m=+1572.173024462" Oct 13 09:10:47 crc kubenswrapper[4685]: I1013 09:10:47.028660 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-r4l79"] Oct 13 09:10:47 crc kubenswrapper[4685]: I1013 09:10:47.037200 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-r4l79"] Oct 13 09:10:47 crc kubenswrapper[4685]: I1013 09:10:47.535645 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86078d92-4584-4ed5-9d3b-280e6678e18d" path="/var/lib/kubelet/pods/86078d92-4584-4ed5-9d3b-280e6678e18d/volumes" Oct 13 09:10:53 crc kubenswrapper[4685]: I1013 09:10:53.050003 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-990a-account-create-8gxck"] Oct 13 09:10:53 crc kubenswrapper[4685]: I1013 09:10:53.068981 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-37cf-account-create-n4sgt"] Oct 13 09:10:53 crc kubenswrapper[4685]: I1013 09:10:53.085990 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-990a-account-create-8gxck"] Oct 13 09:10:53 crc kubenswrapper[4685]: I1013 09:10:53.098849 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-37cf-account-create-n4sgt"] Oct 13 09:10:53 crc kubenswrapper[4685]: I1013 09:10:53.519403 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67ada508-f859-47b2-b17d-009d71fd1af7" path="/var/lib/kubelet/pods/67ada508-f859-47b2-b17d-009d71fd1af7/volumes" Oct 13 09:10:53 crc kubenswrapper[4685]: I1013 09:10:53.523334 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ade16f9d-2076-4f98-ac8b-96e919660fb0" path="/var/lib/kubelet/pods/ade16f9d-2076-4f98-ac8b-96e919660fb0/volumes" Oct 13 09:10:54 crc kubenswrapper[4685]: I1013 09:10:54.322363 4685 scope.go:117] "RemoveContainer" containerID="e5a829666a35fab0d360b6c0caf2d2ff2502b0a900fe7e7811827b1d7f1e2224" Oct 13 09:10:54 crc kubenswrapper[4685]: I1013 09:10:54.346519 4685 scope.go:117] "RemoveContainer" containerID="24b54200b33177414f9bf29827091c56f176ceea6d18f5f169c572ed2d0be71f" Oct 13 09:10:54 crc kubenswrapper[4685]: I1013 09:10:54.374207 4685 scope.go:117] "RemoveContainer" containerID="bf9f33929009619aa40e41abc4290323ad182d516af16469850e9a0ebbefd71b" Oct 13 09:10:54 crc kubenswrapper[4685]: I1013 09:10:54.449377 4685 scope.go:117] "RemoveContainer" containerID="13ac4790a8c78f7ce3c83e12f382c943e0ce69f5e6c37d3f5a09d4d29b11b351" Oct 13 09:10:54 crc kubenswrapper[4685]: I1013 09:10:54.483185 4685 scope.go:117] "RemoveContainer" containerID="34486ea1191b0d949f55df7e09a19546caebcbcc91c252014519208d976df5eb" Oct 13 09:10:54 crc kubenswrapper[4685]: I1013 09:10:54.527754 4685 scope.go:117] "RemoveContainer" containerID="48a7d3f55efb9dfe5ba19f8f906ecb1eb6aa0bb2710440cf37b87eabc8bd6568" Oct 13 09:10:57 crc kubenswrapper[4685]: I1013 09:10:57.048256 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-699e-account-create-8b2g4"] Oct 13 09:10:57 crc kubenswrapper[4685]: I1013 09:10:57.057030 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-699e-account-create-8b2g4"] Oct 13 09:10:57 crc kubenswrapper[4685]: I1013 09:10:57.512794 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24a0a6b1-caf0-4256-b80d-0c6d8171f2de" path="/var/lib/kubelet/pods/24a0a6b1-caf0-4256-b80d-0c6d8171f2de/volumes" Oct 13 09:11:15 crc kubenswrapper[4685]: I1013 09:11:15.037063 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-dp86g"] Oct 13 09:11:15 crc kubenswrapper[4685]: I1013 09:11:15.053957 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-f2mfh"] Oct 13 09:11:15 crc kubenswrapper[4685]: I1013 09:11:15.066066 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-f2mfh"] Oct 13 09:11:15 crc kubenswrapper[4685]: I1013 09:11:15.073072 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-dp86g"] Oct 13 09:11:15 crc kubenswrapper[4685]: I1013 09:11:15.513808 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6870aa8c-c1e3-4c3c-b1dd-c737e0a1d382" path="/var/lib/kubelet/pods/6870aa8c-c1e3-4c3c-b1dd-c737e0a1d382/volumes" Oct 13 09:11:15 crc kubenswrapper[4685]: I1013 09:11:15.514570 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aab3ac63-a140-47f9-9fa3-113d412dd2d5" path="/var/lib/kubelet/pods/aab3ac63-a140-47f9-9fa3-113d412dd2d5/volumes" Oct 13 09:11:20 crc kubenswrapper[4685]: I1013 09:11:20.027437 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-87wjr"] Oct 13 09:11:20 crc kubenswrapper[4685]: I1013 09:11:20.035201 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-87wjr"] Oct 13 09:11:21 crc kubenswrapper[4685]: I1013 09:11:21.518671 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c201a91a-9c59-46e8-bd69-cde6bd87d14a" path="/var/lib/kubelet/pods/c201a91a-9c59-46e8-bd69-cde6bd87d14a/volumes" Oct 13 09:11:25 crc kubenswrapper[4685]: I1013 09:11:25.063416 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-l25fj"] Oct 13 09:11:25 crc kubenswrapper[4685]: I1013 09:11:25.072556 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-l25fj"] Oct 13 09:11:25 crc kubenswrapper[4685]: I1013 09:11:25.555144 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="566d7edf-ec33-43e8-abd4-7e180a2fda57" path="/var/lib/kubelet/pods/566d7edf-ec33-43e8-abd4-7e180a2fda57/volumes" Oct 13 09:11:29 crc kubenswrapper[4685]: I1013 09:11:29.052792 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-jvxxs"] Oct 13 09:11:29 crc kubenswrapper[4685]: I1013 09:11:29.059803 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-jvxxs"] Oct 13 09:11:29 crc kubenswrapper[4685]: I1013 09:11:29.514067 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5a0947d-3515-4e7c-a69f-c7e4d3e9f251" path="/var/lib/kubelet/pods/b5a0947d-3515-4e7c-a69f-c7e4d3e9f251/volumes" Oct 13 09:11:31 crc kubenswrapper[4685]: I1013 09:11:31.025799 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-512d-account-create-nhbjz"] Oct 13 09:11:31 crc kubenswrapper[4685]: I1013 09:11:31.040955 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-512d-account-create-nhbjz"] Oct 13 09:11:31 crc kubenswrapper[4685]: I1013 09:11:31.513198 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="373ea549-fc1c-43eb-8af2-6feb43dd9f70" path="/var/lib/kubelet/pods/373ea549-fc1c-43eb-8af2-6feb43dd9f70/volumes" Oct 13 09:11:32 crc kubenswrapper[4685]: I1013 09:11:32.044805 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-2261-account-create-ndcwb"] Oct 13 09:11:32 crc kubenswrapper[4685]: I1013 09:11:32.069078 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-1a3c-account-create-kqvxn"] Oct 13 09:11:32 crc kubenswrapper[4685]: I1013 09:11:32.078623 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-1a3c-account-create-kqvxn"] Oct 13 09:11:32 crc kubenswrapper[4685]: I1013 09:11:32.087006 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-2261-account-create-ndcwb"] Oct 13 09:11:33 crc kubenswrapper[4685]: I1013 09:11:33.513039 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fe757e4-3653-4673-af9d-cf9895e28d3e" path="/var/lib/kubelet/pods/3fe757e4-3653-4673-af9d-cf9895e28d3e/volumes" Oct 13 09:11:33 crc kubenswrapper[4685]: I1013 09:11:33.513657 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75095121-6e99-4595-b06f-7440639ef8cc" path="/var/lib/kubelet/pods/75095121-6e99-4595-b06f-7440639ef8cc/volumes" Oct 13 09:11:52 crc kubenswrapper[4685]: I1013 09:11:52.980413 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:11:52 crc kubenswrapper[4685]: I1013 09:11:52.980944 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:11:54 crc kubenswrapper[4685]: I1013 09:11:54.669684 4685 scope.go:117] "RemoveContainer" containerID="54524e8771985cb5390931a8625210d02bfffcc3b1af2e7fdb3cb825339da6b6" Oct 13 09:11:54 crc kubenswrapper[4685]: I1013 09:11:54.698046 4685 scope.go:117] "RemoveContainer" containerID="3f7fb51165e65e09f0de8b115d50f44b4dca84584b4d149722bbd704aa2a4718" Oct 13 09:11:54 crc kubenswrapper[4685]: I1013 09:11:54.784412 4685 scope.go:117] "RemoveContainer" containerID="a7425caa7079a13a6b62cc8bbdeb7f3b30af1b8418964a6473386ca4c5d7d855" Oct 13 09:11:54 crc kubenswrapper[4685]: I1013 09:11:54.843735 4685 scope.go:117] "RemoveContainer" containerID="2620c9584b85313a1c833d0dce417aec96153ad94723a334d24b54af9ba9ce20" Oct 13 09:11:54 crc kubenswrapper[4685]: I1013 09:11:54.890012 4685 scope.go:117] "RemoveContainer" containerID="933c94f4464c94619faea471b44761d3cffdd20a36877e83f316467606fbf645" Oct 13 09:11:54 crc kubenswrapper[4685]: I1013 09:11:54.954699 4685 scope.go:117] "RemoveContainer" containerID="6a37957315414d8b9f2fefba688c4bcab605c61cef7b622d0f1955d63cc3d202" Oct 13 09:11:54 crc kubenswrapper[4685]: I1013 09:11:54.998519 4685 scope.go:117] "RemoveContainer" containerID="31e1a7014e1b7c1cec0638f25f223a1b5e50083c2510c2962394510cd25341cb" Oct 13 09:11:55 crc kubenswrapper[4685]: I1013 09:11:55.025543 4685 scope.go:117] "RemoveContainer" containerID="5977da2ef9ebca4c1e0e11708be68c1d3bf8839636a1b5ffde638454afd0c394" Oct 13 09:11:55 crc kubenswrapper[4685]: I1013 09:11:55.045834 4685 scope.go:117] "RemoveContainer" containerID="fc673b5b34bc03076fb56efe1827e2a6d8cc8e9e874ad8f9a0c7d2900481399f" Oct 13 09:11:58 crc kubenswrapper[4685]: I1013 09:11:58.047805 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-4phvd"] Oct 13 09:11:58 crc kubenswrapper[4685]: I1013 09:11:58.066868 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-4phvd"] Oct 13 09:11:59 crc kubenswrapper[4685]: I1013 09:11:59.519447 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5afcbfe-3193-426b-97fe-4e43f4a483d8" path="/var/lib/kubelet/pods/c5afcbfe-3193-426b-97fe-4e43f4a483d8/volumes" Oct 13 09:12:09 crc kubenswrapper[4685]: I1013 09:12:09.960810 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-shff7"] Oct 13 09:12:09 crc kubenswrapper[4685]: I1013 09:12:09.963695 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-shff7" Oct 13 09:12:09 crc kubenswrapper[4685]: I1013 09:12:09.986041 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-shff7"] Oct 13 09:12:10 crc kubenswrapper[4685]: I1013 09:12:10.112216 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b4fb\" (UniqueName: \"kubernetes.io/projected/401f1e68-989c-4b13-9d97-43803807f77f-kube-api-access-9b4fb\") pod \"redhat-marketplace-shff7\" (UID: \"401f1e68-989c-4b13-9d97-43803807f77f\") " pod="openshift-marketplace/redhat-marketplace-shff7" Oct 13 09:12:10 crc kubenswrapper[4685]: I1013 09:12:10.112327 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/401f1e68-989c-4b13-9d97-43803807f77f-catalog-content\") pod \"redhat-marketplace-shff7\" (UID: \"401f1e68-989c-4b13-9d97-43803807f77f\") " pod="openshift-marketplace/redhat-marketplace-shff7" Oct 13 09:12:10 crc kubenswrapper[4685]: I1013 09:12:10.112416 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/401f1e68-989c-4b13-9d97-43803807f77f-utilities\") pod \"redhat-marketplace-shff7\" (UID: \"401f1e68-989c-4b13-9d97-43803807f77f\") " pod="openshift-marketplace/redhat-marketplace-shff7" Oct 13 09:12:10 crc kubenswrapper[4685]: I1013 09:12:10.213854 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/401f1e68-989c-4b13-9d97-43803807f77f-catalog-content\") pod \"redhat-marketplace-shff7\" (UID: \"401f1e68-989c-4b13-9d97-43803807f77f\") " pod="openshift-marketplace/redhat-marketplace-shff7" Oct 13 09:12:10 crc kubenswrapper[4685]: I1013 09:12:10.213990 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/401f1e68-989c-4b13-9d97-43803807f77f-utilities\") pod \"redhat-marketplace-shff7\" (UID: \"401f1e68-989c-4b13-9d97-43803807f77f\") " pod="openshift-marketplace/redhat-marketplace-shff7" Oct 13 09:12:10 crc kubenswrapper[4685]: I1013 09:12:10.214082 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b4fb\" (UniqueName: \"kubernetes.io/projected/401f1e68-989c-4b13-9d97-43803807f77f-kube-api-access-9b4fb\") pod \"redhat-marketplace-shff7\" (UID: \"401f1e68-989c-4b13-9d97-43803807f77f\") " pod="openshift-marketplace/redhat-marketplace-shff7" Oct 13 09:12:10 crc kubenswrapper[4685]: I1013 09:12:10.214572 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/401f1e68-989c-4b13-9d97-43803807f77f-catalog-content\") pod \"redhat-marketplace-shff7\" (UID: \"401f1e68-989c-4b13-9d97-43803807f77f\") " pod="openshift-marketplace/redhat-marketplace-shff7" Oct 13 09:12:10 crc kubenswrapper[4685]: I1013 09:12:10.214593 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/401f1e68-989c-4b13-9d97-43803807f77f-utilities\") pod \"redhat-marketplace-shff7\" (UID: \"401f1e68-989c-4b13-9d97-43803807f77f\") " pod="openshift-marketplace/redhat-marketplace-shff7" Oct 13 09:12:10 crc kubenswrapper[4685]: I1013 09:12:10.236576 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b4fb\" (UniqueName: \"kubernetes.io/projected/401f1e68-989c-4b13-9d97-43803807f77f-kube-api-access-9b4fb\") pod \"redhat-marketplace-shff7\" (UID: \"401f1e68-989c-4b13-9d97-43803807f77f\") " pod="openshift-marketplace/redhat-marketplace-shff7" Oct 13 09:12:10 crc kubenswrapper[4685]: I1013 09:12:10.301556 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-shff7" Oct 13 09:12:10 crc kubenswrapper[4685]: I1013 09:12:10.791346 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-shff7"] Oct 13 09:12:11 crc kubenswrapper[4685]: I1013 09:12:11.664226 4685 generic.go:334] "Generic (PLEG): container finished" podID="401f1e68-989c-4b13-9d97-43803807f77f" containerID="9c6a11de34728c1a3ced3d783b00adbf3fc92bde11f575bc08b81cdba21e8a27" exitCode=0 Oct 13 09:12:11 crc kubenswrapper[4685]: I1013 09:12:11.664282 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-shff7" event={"ID":"401f1e68-989c-4b13-9d97-43803807f77f","Type":"ContainerDied","Data":"9c6a11de34728c1a3ced3d783b00adbf3fc92bde11f575bc08b81cdba21e8a27"} Oct 13 09:12:11 crc kubenswrapper[4685]: I1013 09:12:11.664555 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-shff7" event={"ID":"401f1e68-989c-4b13-9d97-43803807f77f","Type":"ContainerStarted","Data":"e319bd7f447489e53b439a22eb86e7996a23636bd8605a21ea1d46da9ed06f2b"} Oct 13 09:12:12 crc kubenswrapper[4685]: I1013 09:12:12.674337 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-shff7" event={"ID":"401f1e68-989c-4b13-9d97-43803807f77f","Type":"ContainerStarted","Data":"cad1860d6fcfbddf6e2b9948d355462d04c35731f594245ba033109036c3e504"} Oct 13 09:12:13 crc kubenswrapper[4685]: I1013 09:12:13.687687 4685 generic.go:334] "Generic (PLEG): container finished" podID="401f1e68-989c-4b13-9d97-43803807f77f" containerID="cad1860d6fcfbddf6e2b9948d355462d04c35731f594245ba033109036c3e504" exitCode=0 Oct 13 09:12:13 crc kubenswrapper[4685]: I1013 09:12:13.687743 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-shff7" event={"ID":"401f1e68-989c-4b13-9d97-43803807f77f","Type":"ContainerDied","Data":"cad1860d6fcfbddf6e2b9948d355462d04c35731f594245ba033109036c3e504"} Oct 13 09:12:14 crc kubenswrapper[4685]: I1013 09:12:14.704546 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-shff7" event={"ID":"401f1e68-989c-4b13-9d97-43803807f77f","Type":"ContainerStarted","Data":"1a9cf63fe398ff20ae865b007622435d8a4c7a91a3a91cc3cdc9b09686562112"} Oct 13 09:12:14 crc kubenswrapper[4685]: I1013 09:12:14.730317 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-shff7" podStartSLOduration=3.05145367 podStartE2EDuration="5.730297983s" podCreationTimestamp="2025-10-13 09:12:09 +0000 UTC" firstStartedPulling="2025-10-13 09:12:11.666269914 +0000 UTC m=+1656.814145685" lastFinishedPulling="2025-10-13 09:12:14.345114237 +0000 UTC m=+1659.492989998" observedRunningTime="2025-10-13 09:12:14.719782716 +0000 UTC m=+1659.867658477" watchObservedRunningTime="2025-10-13 09:12:14.730297983 +0000 UTC m=+1659.878173744" Oct 13 09:12:16 crc kubenswrapper[4685]: I1013 09:12:16.038794 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-9msgq"] Oct 13 09:12:16 crc kubenswrapper[4685]: I1013 09:12:16.055968 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-9msgq"] Oct 13 09:12:17 crc kubenswrapper[4685]: I1013 09:12:17.524654 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee75d00f-3165-45a5-854d-1045b43f2c87" path="/var/lib/kubelet/pods/ee75d00f-3165-45a5-854d-1045b43f2c87/volumes" Oct 13 09:12:20 crc kubenswrapper[4685]: I1013 09:12:20.302096 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-shff7" Oct 13 09:12:20 crc kubenswrapper[4685]: I1013 09:12:20.302164 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-shff7" Oct 13 09:12:20 crc kubenswrapper[4685]: I1013 09:12:20.368441 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-shff7" Oct 13 09:12:20 crc kubenswrapper[4685]: I1013 09:12:20.768358 4685 generic.go:334] "Generic (PLEG): container finished" podID="49edf7b8-3834-40c0-a038-e6c2388842da" containerID="cacdd8bf2dfff3338c8583d2dc7593c3179d46ad3f96585eae1e55a75337df58" exitCode=0 Oct 13 09:12:20 crc kubenswrapper[4685]: I1013 09:12:20.769181 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc" event={"ID":"49edf7b8-3834-40c0-a038-e6c2388842da","Type":"ContainerDied","Data":"cacdd8bf2dfff3338c8583d2dc7593c3179d46ad3f96585eae1e55a75337df58"} Oct 13 09:12:20 crc kubenswrapper[4685]: I1013 09:12:20.834549 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-shff7" Oct 13 09:12:20 crc kubenswrapper[4685]: I1013 09:12:20.900878 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-shff7"] Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.049982 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-5p6fr"] Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.072651 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-5p6fr"] Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.220095 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc" Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.345337 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/49edf7b8-3834-40c0-a038-e6c2388842da-ssh-key\") pod \"49edf7b8-3834-40c0-a038-e6c2388842da\" (UID: \"49edf7b8-3834-40c0-a038-e6c2388842da\") " Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.345486 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmfnx\" (UniqueName: \"kubernetes.io/projected/49edf7b8-3834-40c0-a038-e6c2388842da-kube-api-access-tmfnx\") pod \"49edf7b8-3834-40c0-a038-e6c2388842da\" (UID: \"49edf7b8-3834-40c0-a038-e6c2388842da\") " Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.345586 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49edf7b8-3834-40c0-a038-e6c2388842da-inventory\") pod \"49edf7b8-3834-40c0-a038-e6c2388842da\" (UID: \"49edf7b8-3834-40c0-a038-e6c2388842da\") " Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.352109 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49edf7b8-3834-40c0-a038-e6c2388842da-kube-api-access-tmfnx" (OuterVolumeSpecName: "kube-api-access-tmfnx") pod "49edf7b8-3834-40c0-a038-e6c2388842da" (UID: "49edf7b8-3834-40c0-a038-e6c2388842da"). InnerVolumeSpecName "kube-api-access-tmfnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.373707 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49edf7b8-3834-40c0-a038-e6c2388842da-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "49edf7b8-3834-40c0-a038-e6c2388842da" (UID: "49edf7b8-3834-40c0-a038-e6c2388842da"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.381565 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49edf7b8-3834-40c0-a038-e6c2388842da-inventory" (OuterVolumeSpecName: "inventory") pod "49edf7b8-3834-40c0-a038-e6c2388842da" (UID: "49edf7b8-3834-40c0-a038-e6c2388842da"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.448119 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/49edf7b8-3834-40c0-a038-e6c2388842da-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.448168 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmfnx\" (UniqueName: \"kubernetes.io/projected/49edf7b8-3834-40c0-a038-e6c2388842da-kube-api-access-tmfnx\") on node \"crc\" DevicePath \"\"" Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.448188 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49edf7b8-3834-40c0-a038-e6c2388842da-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.787392 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc" event={"ID":"49edf7b8-3834-40c0-a038-e6c2388842da","Type":"ContainerDied","Data":"e70170c92131416a3c540af44018baa4a452f6b1d7d8d86f3672ad3bb8d01d5f"} Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.787431 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e70170c92131416a3c540af44018baa4a452f6b1d7d8d86f3672ad3bb8d01d5f" Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.787406 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc" Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.787507 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-shff7" podUID="401f1e68-989c-4b13-9d97-43803807f77f" containerName="registry-server" containerID="cri-o://1a9cf63fe398ff20ae865b007622435d8a4c7a91a3a91cc3cdc9b09686562112" gracePeriod=2 Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.880752 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs"] Oct 13 09:12:22 crc kubenswrapper[4685]: E1013 09:12:22.881128 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49edf7b8-3834-40c0-a038-e6c2388842da" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.881145 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="49edf7b8-3834-40c0-a038-e6c2388842da" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.881368 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="49edf7b8-3834-40c0-a038-e6c2388842da" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.882006 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs" Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.892399 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.892474 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hcn4z" Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.892605 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.894046 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.908816 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs"] Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.958517 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5d774ea-039f-4303-b889-8a6c79077453-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs\" (UID: \"f5d774ea-039f-4303-b889-8a6c79077453\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs" Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.958615 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8726s\" (UniqueName: \"kubernetes.io/projected/f5d774ea-039f-4303-b889-8a6c79077453-kube-api-access-8726s\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs\" (UID: \"f5d774ea-039f-4303-b889-8a6c79077453\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs" Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.958651 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5d774ea-039f-4303-b889-8a6c79077453-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs\" (UID: \"f5d774ea-039f-4303-b889-8a6c79077453\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs" Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.980085 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:12:22 crc kubenswrapper[4685]: I1013 09:12:22.980137 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.060606 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5d774ea-039f-4303-b889-8a6c79077453-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs\" (UID: \"f5d774ea-039f-4303-b889-8a6c79077453\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.060726 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8726s\" (UniqueName: \"kubernetes.io/projected/f5d774ea-039f-4303-b889-8a6c79077453-kube-api-access-8726s\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs\" (UID: \"f5d774ea-039f-4303-b889-8a6c79077453\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.061149 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5d774ea-039f-4303-b889-8a6c79077453-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs\" (UID: \"f5d774ea-039f-4303-b889-8a6c79077453\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.067894 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5d774ea-039f-4303-b889-8a6c79077453-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs\" (UID: \"f5d774ea-039f-4303-b889-8a6c79077453\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.068187 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5d774ea-039f-4303-b889-8a6c79077453-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs\" (UID: \"f5d774ea-039f-4303-b889-8a6c79077453\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.088053 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8726s\" (UniqueName: \"kubernetes.io/projected/f5d774ea-039f-4303-b889-8a6c79077453-kube-api-access-8726s\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs\" (UID: \"f5d774ea-039f-4303-b889-8a6c79077453\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.222733 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-shff7" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.263446 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.263651 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9b4fb\" (UniqueName: \"kubernetes.io/projected/401f1e68-989c-4b13-9d97-43803807f77f-kube-api-access-9b4fb\") pod \"401f1e68-989c-4b13-9d97-43803807f77f\" (UID: \"401f1e68-989c-4b13-9d97-43803807f77f\") " Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.263826 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/401f1e68-989c-4b13-9d97-43803807f77f-utilities\") pod \"401f1e68-989c-4b13-9d97-43803807f77f\" (UID: \"401f1e68-989c-4b13-9d97-43803807f77f\") " Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.263886 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/401f1e68-989c-4b13-9d97-43803807f77f-catalog-content\") pod \"401f1e68-989c-4b13-9d97-43803807f77f\" (UID: \"401f1e68-989c-4b13-9d97-43803807f77f\") " Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.270615 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/401f1e68-989c-4b13-9d97-43803807f77f-utilities" (OuterVolumeSpecName: "utilities") pod "401f1e68-989c-4b13-9d97-43803807f77f" (UID: "401f1e68-989c-4b13-9d97-43803807f77f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.273136 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/401f1e68-989c-4b13-9d97-43803807f77f-kube-api-access-9b4fb" (OuterVolumeSpecName: "kube-api-access-9b4fb") pod "401f1e68-989c-4b13-9d97-43803807f77f" (UID: "401f1e68-989c-4b13-9d97-43803807f77f"). InnerVolumeSpecName "kube-api-access-9b4fb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.283702 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/401f1e68-989c-4b13-9d97-43803807f77f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "401f1e68-989c-4b13-9d97-43803807f77f" (UID: "401f1e68-989c-4b13-9d97-43803807f77f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.366519 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/401f1e68-989c-4b13-9d97-43803807f77f-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.366544 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/401f1e68-989c-4b13-9d97-43803807f77f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.366556 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9b4fb\" (UniqueName: \"kubernetes.io/projected/401f1e68-989c-4b13-9d97-43803807f77f-kube-api-access-9b4fb\") on node \"crc\" DevicePath \"\"" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.516457 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05506d41-5449-4eb6-86ce-0ac5f5c880ba" path="/var/lib/kubelet/pods/05506d41-5449-4eb6-86ce-0ac5f5c880ba/volumes" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.805573 4685 generic.go:334] "Generic (PLEG): container finished" podID="401f1e68-989c-4b13-9d97-43803807f77f" containerID="1a9cf63fe398ff20ae865b007622435d8a4c7a91a3a91cc3cdc9b09686562112" exitCode=0 Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.805632 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-shff7" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.805650 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-shff7" event={"ID":"401f1e68-989c-4b13-9d97-43803807f77f","Type":"ContainerDied","Data":"1a9cf63fe398ff20ae865b007622435d8a4c7a91a3a91cc3cdc9b09686562112"} Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.805754 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-shff7" event={"ID":"401f1e68-989c-4b13-9d97-43803807f77f","Type":"ContainerDied","Data":"e319bd7f447489e53b439a22eb86e7996a23636bd8605a21ea1d46da9ed06f2b"} Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.805815 4685 scope.go:117] "RemoveContainer" containerID="1a9cf63fe398ff20ae865b007622435d8a4c7a91a3a91cc3cdc9b09686562112" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.830031 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs"] Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.850343 4685 scope.go:117] "RemoveContainer" containerID="cad1860d6fcfbddf6e2b9948d355462d04c35731f594245ba033109036c3e504" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.859357 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-shff7"] Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.870462 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-shff7"] Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.917561 4685 scope.go:117] "RemoveContainer" containerID="9c6a11de34728c1a3ced3d783b00adbf3fc92bde11f575bc08b81cdba21e8a27" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.933715 4685 scope.go:117] "RemoveContainer" containerID="1a9cf63fe398ff20ae865b007622435d8a4c7a91a3a91cc3cdc9b09686562112" Oct 13 09:12:23 crc kubenswrapper[4685]: E1013 09:12:23.934282 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a9cf63fe398ff20ae865b007622435d8a4c7a91a3a91cc3cdc9b09686562112\": container with ID starting with 1a9cf63fe398ff20ae865b007622435d8a4c7a91a3a91cc3cdc9b09686562112 not found: ID does not exist" containerID="1a9cf63fe398ff20ae865b007622435d8a4c7a91a3a91cc3cdc9b09686562112" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.934356 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a9cf63fe398ff20ae865b007622435d8a4c7a91a3a91cc3cdc9b09686562112"} err="failed to get container status \"1a9cf63fe398ff20ae865b007622435d8a4c7a91a3a91cc3cdc9b09686562112\": rpc error: code = NotFound desc = could not find container \"1a9cf63fe398ff20ae865b007622435d8a4c7a91a3a91cc3cdc9b09686562112\": container with ID starting with 1a9cf63fe398ff20ae865b007622435d8a4c7a91a3a91cc3cdc9b09686562112 not found: ID does not exist" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.934394 4685 scope.go:117] "RemoveContainer" containerID="cad1860d6fcfbddf6e2b9948d355462d04c35731f594245ba033109036c3e504" Oct 13 09:12:23 crc kubenswrapper[4685]: E1013 09:12:23.934845 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cad1860d6fcfbddf6e2b9948d355462d04c35731f594245ba033109036c3e504\": container with ID starting with cad1860d6fcfbddf6e2b9948d355462d04c35731f594245ba033109036c3e504 not found: ID does not exist" containerID="cad1860d6fcfbddf6e2b9948d355462d04c35731f594245ba033109036c3e504" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.934897 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cad1860d6fcfbddf6e2b9948d355462d04c35731f594245ba033109036c3e504"} err="failed to get container status \"cad1860d6fcfbddf6e2b9948d355462d04c35731f594245ba033109036c3e504\": rpc error: code = NotFound desc = could not find container \"cad1860d6fcfbddf6e2b9948d355462d04c35731f594245ba033109036c3e504\": container with ID starting with cad1860d6fcfbddf6e2b9948d355462d04c35731f594245ba033109036c3e504 not found: ID does not exist" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.934942 4685 scope.go:117] "RemoveContainer" containerID="9c6a11de34728c1a3ced3d783b00adbf3fc92bde11f575bc08b81cdba21e8a27" Oct 13 09:12:23 crc kubenswrapper[4685]: E1013 09:12:23.935437 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c6a11de34728c1a3ced3d783b00adbf3fc92bde11f575bc08b81cdba21e8a27\": container with ID starting with 9c6a11de34728c1a3ced3d783b00adbf3fc92bde11f575bc08b81cdba21e8a27 not found: ID does not exist" containerID="9c6a11de34728c1a3ced3d783b00adbf3fc92bde11f575bc08b81cdba21e8a27" Oct 13 09:12:23 crc kubenswrapper[4685]: I1013 09:12:23.935560 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c6a11de34728c1a3ced3d783b00adbf3fc92bde11f575bc08b81cdba21e8a27"} err="failed to get container status \"9c6a11de34728c1a3ced3d783b00adbf3fc92bde11f575bc08b81cdba21e8a27\": rpc error: code = NotFound desc = could not find container \"9c6a11de34728c1a3ced3d783b00adbf3fc92bde11f575bc08b81cdba21e8a27\": container with ID starting with 9c6a11de34728c1a3ced3d783b00adbf3fc92bde11f575bc08b81cdba21e8a27 not found: ID does not exist" Oct 13 09:12:24 crc kubenswrapper[4685]: I1013 09:12:24.819798 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs" event={"ID":"f5d774ea-039f-4303-b889-8a6c79077453","Type":"ContainerStarted","Data":"f31b39ec0e927cdd65bca23ed37fae473973f9636c0698722e324b9e83556b35"} Oct 13 09:12:24 crc kubenswrapper[4685]: I1013 09:12:24.820220 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs" event={"ID":"f5d774ea-039f-4303-b889-8a6c79077453","Type":"ContainerStarted","Data":"6c1507be3205cd6081aaeba6cb47671807ede8a2d31f2c2cf305e3a064455515"} Oct 13 09:12:24 crc kubenswrapper[4685]: I1013 09:12:24.843243 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs" podStartSLOduration=2.314467474 podStartE2EDuration="2.84322543s" podCreationTimestamp="2025-10-13 09:12:22 +0000 UTC" firstStartedPulling="2025-10-13 09:12:23.86803106 +0000 UTC m=+1669.015906821" lastFinishedPulling="2025-10-13 09:12:24.396788976 +0000 UTC m=+1669.544664777" observedRunningTime="2025-10-13 09:12:24.836864677 +0000 UTC m=+1669.984740438" watchObservedRunningTime="2025-10-13 09:12:24.84322543 +0000 UTC m=+1669.991101191" Oct 13 09:12:25 crc kubenswrapper[4685]: I1013 09:12:25.527498 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="401f1e68-989c-4b13-9d97-43803807f77f" path="/var/lib/kubelet/pods/401f1e68-989c-4b13-9d97-43803807f77f/volumes" Oct 13 09:12:29 crc kubenswrapper[4685]: I1013 09:12:29.046573 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-b29nc"] Oct 13 09:12:29 crc kubenswrapper[4685]: I1013 09:12:29.053358 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-b29nc"] Oct 13 09:12:29 crc kubenswrapper[4685]: I1013 09:12:29.523623 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40fbc731-ab15-4cc0-9eab-1ec6c0848cbc" path="/var/lib/kubelet/pods/40fbc731-ab15-4cc0-9eab-1ec6c0848cbc/volumes" Oct 13 09:12:45 crc kubenswrapper[4685]: I1013 09:12:45.043615 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-7h6qb"] Oct 13 09:12:45 crc kubenswrapper[4685]: I1013 09:12:45.057812 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-7h6qb"] Oct 13 09:12:45 crc kubenswrapper[4685]: I1013 09:12:45.518165 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b225d240-40a0-4d59-aa19-6355515e9f8c" path="/var/lib/kubelet/pods/b225d240-40a0-4d59-aa19-6355515e9f8c/volumes" Oct 13 09:12:52 crc kubenswrapper[4685]: I1013 09:12:52.980543 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:12:52 crc kubenswrapper[4685]: I1013 09:12:52.980935 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:12:52 crc kubenswrapper[4685]: I1013 09:12:52.980982 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 09:12:52 crc kubenswrapper[4685]: I1013 09:12:52.981858 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570"} pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 09:12:52 crc kubenswrapper[4685]: I1013 09:12:52.981932 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" containerID="cri-o://199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" gracePeriod=600 Oct 13 09:12:53 crc kubenswrapper[4685]: E1013 09:12:53.106269 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:12:54 crc kubenswrapper[4685]: I1013 09:12:54.095817 4685 generic.go:334] "Generic (PLEG): container finished" podID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" exitCode=0 Oct 13 09:12:54 crc kubenswrapper[4685]: I1013 09:12:54.095889 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerDied","Data":"199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570"} Oct 13 09:12:54 crc kubenswrapper[4685]: I1013 09:12:54.096195 4685 scope.go:117] "RemoveContainer" containerID="a2277b2f3c48d695283091e6c5552e7453d3c307f3bc3a11a7a1f02c1122c7c2" Oct 13 09:12:54 crc kubenswrapper[4685]: I1013 09:12:54.097306 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:12:54 crc kubenswrapper[4685]: E1013 09:12:54.097802 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:12:55 crc kubenswrapper[4685]: I1013 09:12:55.241367 4685 scope.go:117] "RemoveContainer" containerID="048241ab2b6e54a12029b77e838f010d8527358e132afb5585f66a15de926667" Oct 13 09:12:55 crc kubenswrapper[4685]: I1013 09:12:55.276327 4685 scope.go:117] "RemoveContainer" containerID="ee542b97c2ea580a1848390372ad4dd4d87e49fcb174c14691b7a9017f8a1a5c" Oct 13 09:12:55 crc kubenswrapper[4685]: I1013 09:12:55.317378 4685 scope.go:117] "RemoveContainer" containerID="114f444907b28763ac128396f5487639845624f6c74051d43cbdffb59c484704" Oct 13 09:12:55 crc kubenswrapper[4685]: I1013 09:12:55.364744 4685 scope.go:117] "RemoveContainer" containerID="16c5d0f261753dbe0cc387ffa98a76a6a4a86536f143165e934db71b95f53192" Oct 13 09:12:55 crc kubenswrapper[4685]: I1013 09:12:55.416675 4685 scope.go:117] "RemoveContainer" containerID="d9b0e879150b37dc976a9e934abaa9d8257375d68d2c2ac8dc3ac813e49f3cd8" Oct 13 09:13:04 crc kubenswrapper[4685]: I1013 09:13:04.503284 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:13:04 crc kubenswrapper[4685]: E1013 09:13:04.504006 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:13:11 crc kubenswrapper[4685]: I1013 09:13:11.251175 4685 generic.go:334] "Generic (PLEG): container finished" podID="b001d17a-1aea-44ba-86c5-ba6b312156c1" containerID="f4bd2015aa86f0bf7c40324d231108b8eefc9288b86f37bc304379cab4fbde84" exitCode=1 Oct 13 09:13:11 crc kubenswrapper[4685]: I1013 09:13:11.251232 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" event={"ID":"b001d17a-1aea-44ba-86c5-ba6b312156c1","Type":"ContainerDied","Data":"f4bd2015aa86f0bf7c40324d231108b8eefc9288b86f37bc304379cab4fbde84"} Oct 13 09:13:11 crc kubenswrapper[4685]: I1013 09:13:11.251786 4685 scope.go:117] "RemoveContainer" containerID="cedf68f34af27ba6b0e7b6a3ebc8c368caa3b020eacec7a71ecf4f542d94655f" Oct 13 09:13:11 crc kubenswrapper[4685]: I1013 09:13:11.252602 4685 scope.go:117] "RemoveContainer" containerID="f4bd2015aa86f0bf7c40324d231108b8eefc9288b86f37bc304379cab4fbde84" Oct 13 09:13:11 crc kubenswrapper[4685]: E1013 09:13:11.252963 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:13:13 crc kubenswrapper[4685]: I1013 09:13:13.442723 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:13:13 crc kubenswrapper[4685]: I1013 09:13:13.444308 4685 scope.go:117] "RemoveContainer" containerID="f4bd2015aa86f0bf7c40324d231108b8eefc9288b86f37bc304379cab4fbde84" Oct 13 09:13:13 crc kubenswrapper[4685]: E1013 09:13:13.444698 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:13:15 crc kubenswrapper[4685]: I1013 09:13:15.514789 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:13:15 crc kubenswrapper[4685]: E1013 09:13:15.515484 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:13:22 crc kubenswrapper[4685]: I1013 09:13:22.045121 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-kxkml"] Oct 13 09:13:22 crc kubenswrapper[4685]: I1013 09:13:22.052951 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-kxkml"] Oct 13 09:13:23 crc kubenswrapper[4685]: I1013 09:13:23.038752 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-sjt68"] Oct 13 09:13:23 crc kubenswrapper[4685]: I1013 09:13:23.053852 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-rtvbl"] Oct 13 09:13:23 crc kubenswrapper[4685]: I1013 09:13:23.063422 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-rtvbl"] Oct 13 09:13:23 crc kubenswrapper[4685]: I1013 09:13:23.071055 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-sjt68"] Oct 13 09:13:23 crc kubenswrapper[4685]: I1013 09:13:23.442967 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:13:23 crc kubenswrapper[4685]: I1013 09:13:23.444084 4685 scope.go:117] "RemoveContainer" containerID="f4bd2015aa86f0bf7c40324d231108b8eefc9288b86f37bc304379cab4fbde84" Oct 13 09:13:23 crc kubenswrapper[4685]: E1013 09:13:23.444526 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:13:23 crc kubenswrapper[4685]: I1013 09:13:23.518592 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4119a02b-4484-438b-8521-33332cd63834" path="/var/lib/kubelet/pods/4119a02b-4484-438b-8521-33332cd63834/volumes" Oct 13 09:13:23 crc kubenswrapper[4685]: I1013 09:13:23.519459 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8826d2db-c766-469a-b840-83e177ebf61f" path="/var/lib/kubelet/pods/8826d2db-c766-469a-b840-83e177ebf61f/volumes" Oct 13 09:13:23 crc kubenswrapper[4685]: I1013 09:13:23.520270 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da42d77d-8d37-4e97-8852-2976d0be3012" path="/var/lib/kubelet/pods/da42d77d-8d37-4e97-8852-2976d0be3012/volumes" Oct 13 09:13:28 crc kubenswrapper[4685]: I1013 09:13:28.502561 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:13:28 crc kubenswrapper[4685]: E1013 09:13:28.502839 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:13:33 crc kubenswrapper[4685]: I1013 09:13:33.068283 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-cf2b-account-create-d9lnd"] Oct 13 09:13:33 crc kubenswrapper[4685]: I1013 09:13:33.078988 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-379c-account-create-z5kgm"] Oct 13 09:13:33 crc kubenswrapper[4685]: I1013 09:13:33.093539 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-cf2b-account-create-d9lnd"] Oct 13 09:13:33 crc kubenswrapper[4685]: I1013 09:13:33.102155 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-379c-account-create-z5kgm"] Oct 13 09:13:33 crc kubenswrapper[4685]: I1013 09:13:33.524153 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2581060-597a-46b8-861f-ff53d55748fc" path="/var/lib/kubelet/pods/a2581060-597a-46b8-861f-ff53d55748fc/volumes" Oct 13 09:13:33 crc kubenswrapper[4685]: I1013 09:13:33.524683 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe673980-2c45-45d0-a724-0b66c6f3e5ee" path="/var/lib/kubelet/pods/fe673980-2c45-45d0-a724-0b66c6f3e5ee/volumes" Oct 13 09:13:34 crc kubenswrapper[4685]: I1013 09:13:34.031120 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-b46c-account-create-2rssp"] Oct 13 09:13:34 crc kubenswrapper[4685]: I1013 09:13:34.037482 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-b46c-account-create-2rssp"] Oct 13 09:13:35 crc kubenswrapper[4685]: I1013 09:13:35.520571 4685 scope.go:117] "RemoveContainer" containerID="f4bd2015aa86f0bf7c40324d231108b8eefc9288b86f37bc304379cab4fbde84" Oct 13 09:13:35 crc kubenswrapper[4685]: E1013 09:13:35.521114 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:13:35 crc kubenswrapper[4685]: I1013 09:13:35.522140 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="998d6585-7a36-44b8-a339-a7c8409df426" path="/var/lib/kubelet/pods/998d6585-7a36-44b8-a339-a7c8409df426/volumes" Oct 13 09:13:37 crc kubenswrapper[4685]: I1013 09:13:37.492708 4685 generic.go:334] "Generic (PLEG): container finished" podID="f5d774ea-039f-4303-b889-8a6c79077453" containerID="f31b39ec0e927cdd65bca23ed37fae473973f9636c0698722e324b9e83556b35" exitCode=0 Oct 13 09:13:37 crc kubenswrapper[4685]: I1013 09:13:37.492793 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs" event={"ID":"f5d774ea-039f-4303-b889-8a6c79077453","Type":"ContainerDied","Data":"f31b39ec0e927cdd65bca23ed37fae473973f9636c0698722e324b9e83556b35"} Oct 13 09:13:38 crc kubenswrapper[4685]: I1013 09:13:38.928999 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs" Oct 13 09:13:38 crc kubenswrapper[4685]: I1013 09:13:38.960388 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5d774ea-039f-4303-b889-8a6c79077453-inventory\") pod \"f5d774ea-039f-4303-b889-8a6c79077453\" (UID: \"f5d774ea-039f-4303-b889-8a6c79077453\") " Oct 13 09:13:38 crc kubenswrapper[4685]: I1013 09:13:38.960457 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5d774ea-039f-4303-b889-8a6c79077453-ssh-key\") pod \"f5d774ea-039f-4303-b889-8a6c79077453\" (UID: \"f5d774ea-039f-4303-b889-8a6c79077453\") " Oct 13 09:13:38 crc kubenswrapper[4685]: I1013 09:13:38.960503 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8726s\" (UniqueName: \"kubernetes.io/projected/f5d774ea-039f-4303-b889-8a6c79077453-kube-api-access-8726s\") pod \"f5d774ea-039f-4303-b889-8a6c79077453\" (UID: \"f5d774ea-039f-4303-b889-8a6c79077453\") " Oct 13 09:13:38 crc kubenswrapper[4685]: I1013 09:13:38.966212 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5d774ea-039f-4303-b889-8a6c79077453-kube-api-access-8726s" (OuterVolumeSpecName: "kube-api-access-8726s") pod "f5d774ea-039f-4303-b889-8a6c79077453" (UID: "f5d774ea-039f-4303-b889-8a6c79077453"). InnerVolumeSpecName "kube-api-access-8726s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:13:38 crc kubenswrapper[4685]: I1013 09:13:38.990443 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5d774ea-039f-4303-b889-8a6c79077453-inventory" (OuterVolumeSpecName: "inventory") pod "f5d774ea-039f-4303-b889-8a6c79077453" (UID: "f5d774ea-039f-4303-b889-8a6c79077453"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:13:38 crc kubenswrapper[4685]: I1013 09:13:38.994372 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5d774ea-039f-4303-b889-8a6c79077453-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f5d774ea-039f-4303-b889-8a6c79077453" (UID: "f5d774ea-039f-4303-b889-8a6c79077453"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.062931 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5d774ea-039f-4303-b889-8a6c79077453-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.063264 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5d774ea-039f-4303-b889-8a6c79077453-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.063280 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8726s\" (UniqueName: \"kubernetes.io/projected/f5d774ea-039f-4303-b889-8a6c79077453-kube-api-access-8726s\") on node \"crc\" DevicePath \"\"" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.534399 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs" event={"ID":"f5d774ea-039f-4303-b889-8a6c79077453","Type":"ContainerDied","Data":"6c1507be3205cd6081aaeba6cb47671807ede8a2d31f2c2cf305e3a064455515"} Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.534469 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c1507be3205cd6081aaeba6cb47671807ede8a2d31f2c2cf305e3a064455515" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.534628 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.616749 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz"] Oct 13 09:13:39 crc kubenswrapper[4685]: E1013 09:13:39.617133 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401f1e68-989c-4b13-9d97-43803807f77f" containerName="registry-server" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.617154 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="401f1e68-989c-4b13-9d97-43803807f77f" containerName="registry-server" Oct 13 09:13:39 crc kubenswrapper[4685]: E1013 09:13:39.617185 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d774ea-039f-4303-b889-8a6c79077453" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.617195 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d774ea-039f-4303-b889-8a6c79077453" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 13 09:13:39 crc kubenswrapper[4685]: E1013 09:13:39.617207 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401f1e68-989c-4b13-9d97-43803807f77f" containerName="extract-utilities" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.617215 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="401f1e68-989c-4b13-9d97-43803807f77f" containerName="extract-utilities" Oct 13 09:13:39 crc kubenswrapper[4685]: E1013 09:13:39.617223 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401f1e68-989c-4b13-9d97-43803807f77f" containerName="extract-content" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.617228 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="401f1e68-989c-4b13-9d97-43803807f77f" containerName="extract-content" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.617399 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5d774ea-039f-4303-b889-8a6c79077453" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.617441 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="401f1e68-989c-4b13-9d97-43803807f77f" containerName="registry-server" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.618078 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.621842 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.622096 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hcn4z" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.622211 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.622312 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.664907 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz"] Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.678033 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6535c66b-9311-4170-8fa5-c3d79b5cd7af-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz\" (UID: \"6535c66b-9311-4170-8fa5-c3d79b5cd7af\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.678079 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6535c66b-9311-4170-8fa5-c3d79b5cd7af-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz\" (UID: \"6535c66b-9311-4170-8fa5-c3d79b5cd7af\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.678100 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c74p9\" (UniqueName: \"kubernetes.io/projected/6535c66b-9311-4170-8fa5-c3d79b5cd7af-kube-api-access-c74p9\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz\" (UID: \"6535c66b-9311-4170-8fa5-c3d79b5cd7af\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.779803 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6535c66b-9311-4170-8fa5-c3d79b5cd7af-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz\" (UID: \"6535c66b-9311-4170-8fa5-c3d79b5cd7af\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.779848 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6535c66b-9311-4170-8fa5-c3d79b5cd7af-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz\" (UID: \"6535c66b-9311-4170-8fa5-c3d79b5cd7af\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.779873 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c74p9\" (UniqueName: \"kubernetes.io/projected/6535c66b-9311-4170-8fa5-c3d79b5cd7af-kube-api-access-c74p9\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz\" (UID: \"6535c66b-9311-4170-8fa5-c3d79b5cd7af\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.788735 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6535c66b-9311-4170-8fa5-c3d79b5cd7af-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz\" (UID: \"6535c66b-9311-4170-8fa5-c3d79b5cd7af\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.788744 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6535c66b-9311-4170-8fa5-c3d79b5cd7af-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz\" (UID: \"6535c66b-9311-4170-8fa5-c3d79b5cd7af\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.800637 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c74p9\" (UniqueName: \"kubernetes.io/projected/6535c66b-9311-4170-8fa5-c3d79b5cd7af-kube-api-access-c74p9\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz\" (UID: \"6535c66b-9311-4170-8fa5-c3d79b5cd7af\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz" Oct 13 09:13:39 crc kubenswrapper[4685]: I1013 09:13:39.932616 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz" Oct 13 09:13:40 crc kubenswrapper[4685]: I1013 09:13:40.666266 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz"] Oct 13 09:13:41 crc kubenswrapper[4685]: I1013 09:13:41.553297 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz" event={"ID":"6535c66b-9311-4170-8fa5-c3d79b5cd7af","Type":"ContainerStarted","Data":"717e58dd51b266e3469fa4a2f8238d12fd63d0fca58adc22aec8b9b53649025f"} Oct 13 09:13:41 crc kubenswrapper[4685]: I1013 09:13:41.553937 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz" event={"ID":"6535c66b-9311-4170-8fa5-c3d79b5cd7af","Type":"ContainerStarted","Data":"2ba391ddf977e9dcd857f6dab238f4fe33ed44d6cc22ad0bf65c386bce1ff14f"} Oct 13 09:13:41 crc kubenswrapper[4685]: I1013 09:13:41.575021 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz" podStartSLOduration=2.058389432 podStartE2EDuration="2.575005397s" podCreationTimestamp="2025-10-13 09:13:39 +0000 UTC" firstStartedPulling="2025-10-13 09:13:40.681887691 +0000 UTC m=+1745.829763442" lastFinishedPulling="2025-10-13 09:13:41.198503646 +0000 UTC m=+1746.346379407" observedRunningTime="2025-10-13 09:13:41.569970199 +0000 UTC m=+1746.717845970" watchObservedRunningTime="2025-10-13 09:13:41.575005397 +0000 UTC m=+1746.722881158" Oct 13 09:13:43 crc kubenswrapper[4685]: I1013 09:13:43.502868 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:13:43 crc kubenswrapper[4685]: E1013 09:13:43.503539 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:13:46 crc kubenswrapper[4685]: I1013 09:13:46.502256 4685 scope.go:117] "RemoveContainer" containerID="f4bd2015aa86f0bf7c40324d231108b8eefc9288b86f37bc304379cab4fbde84" Oct 13 09:13:46 crc kubenswrapper[4685]: E1013 09:13:46.502787 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:13:46 crc kubenswrapper[4685]: I1013 09:13:46.596800 4685 generic.go:334] "Generic (PLEG): container finished" podID="6535c66b-9311-4170-8fa5-c3d79b5cd7af" containerID="717e58dd51b266e3469fa4a2f8238d12fd63d0fca58adc22aec8b9b53649025f" exitCode=0 Oct 13 09:13:46 crc kubenswrapper[4685]: I1013 09:13:46.596836 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz" event={"ID":"6535c66b-9311-4170-8fa5-c3d79b5cd7af","Type":"ContainerDied","Data":"717e58dd51b266e3469fa4a2f8238d12fd63d0fca58adc22aec8b9b53649025f"} Oct 13 09:13:47 crc kubenswrapper[4685]: I1013 09:13:47.996205 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.051577 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c74p9\" (UniqueName: \"kubernetes.io/projected/6535c66b-9311-4170-8fa5-c3d79b5cd7af-kube-api-access-c74p9\") pod \"6535c66b-9311-4170-8fa5-c3d79b5cd7af\" (UID: \"6535c66b-9311-4170-8fa5-c3d79b5cd7af\") " Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.051786 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6535c66b-9311-4170-8fa5-c3d79b5cd7af-inventory\") pod \"6535c66b-9311-4170-8fa5-c3d79b5cd7af\" (UID: \"6535c66b-9311-4170-8fa5-c3d79b5cd7af\") " Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.051942 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6535c66b-9311-4170-8fa5-c3d79b5cd7af-ssh-key\") pod \"6535c66b-9311-4170-8fa5-c3d79b5cd7af\" (UID: \"6535c66b-9311-4170-8fa5-c3d79b5cd7af\") " Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.069225 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6535c66b-9311-4170-8fa5-c3d79b5cd7af-kube-api-access-c74p9" (OuterVolumeSpecName: "kube-api-access-c74p9") pod "6535c66b-9311-4170-8fa5-c3d79b5cd7af" (UID: "6535c66b-9311-4170-8fa5-c3d79b5cd7af"). InnerVolumeSpecName "kube-api-access-c74p9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.087398 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6535c66b-9311-4170-8fa5-c3d79b5cd7af-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6535c66b-9311-4170-8fa5-c3d79b5cd7af" (UID: "6535c66b-9311-4170-8fa5-c3d79b5cd7af"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.097564 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6535c66b-9311-4170-8fa5-c3d79b5cd7af-inventory" (OuterVolumeSpecName: "inventory") pod "6535c66b-9311-4170-8fa5-c3d79b5cd7af" (UID: "6535c66b-9311-4170-8fa5-c3d79b5cd7af"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.155330 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6535c66b-9311-4170-8fa5-c3d79b5cd7af-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.155612 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6535c66b-9311-4170-8fa5-c3d79b5cd7af-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.155671 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c74p9\" (UniqueName: \"kubernetes.io/projected/6535c66b-9311-4170-8fa5-c3d79b5cd7af-kube-api-access-c74p9\") on node \"crc\" DevicePath \"\"" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.618882 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz" event={"ID":"6535c66b-9311-4170-8fa5-c3d79b5cd7af","Type":"ContainerDied","Data":"2ba391ddf977e9dcd857f6dab238f4fe33ed44d6cc22ad0bf65c386bce1ff14f"} Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.618954 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ba391ddf977e9dcd857f6dab238f4fe33ed44d6cc22ad0bf65c386bce1ff14f" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.619001 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.689498 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-s2r7z"] Oct 13 09:13:48 crc kubenswrapper[4685]: E1013 09:13:48.689883 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6535c66b-9311-4170-8fa5-c3d79b5cd7af" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.689923 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="6535c66b-9311-4170-8fa5-c3d79b5cd7af" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.690086 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="6535c66b-9311-4170-8fa5-c3d79b5cd7af" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.690705 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-s2r7z" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.695574 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.695610 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.696494 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.696668 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hcn4z" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.708417 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-s2r7z"] Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.767931 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-s2r7z\" (UID: \"5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-s2r7z" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.768076 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-s2r7z\" (UID: \"5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-s2r7z" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.768120 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skk58\" (UniqueName: \"kubernetes.io/projected/5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9-kube-api-access-skk58\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-s2r7z\" (UID: \"5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-s2r7z" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.869588 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-s2r7z\" (UID: \"5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-s2r7z" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.869661 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skk58\" (UniqueName: \"kubernetes.io/projected/5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9-kube-api-access-skk58\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-s2r7z\" (UID: \"5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-s2r7z" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.869743 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-s2r7z\" (UID: \"5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-s2r7z" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.874187 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-s2r7z\" (UID: \"5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-s2r7z" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.874477 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-s2r7z\" (UID: \"5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-s2r7z" Oct 13 09:13:48 crc kubenswrapper[4685]: I1013 09:13:48.884610 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skk58\" (UniqueName: \"kubernetes.io/projected/5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9-kube-api-access-skk58\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-s2r7z\" (UID: \"5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-s2r7z" Oct 13 09:13:49 crc kubenswrapper[4685]: I1013 09:13:49.010302 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-s2r7z" Oct 13 09:13:49 crc kubenswrapper[4685]: I1013 09:13:49.524967 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-s2r7z"] Oct 13 09:13:49 crc kubenswrapper[4685]: I1013 09:13:49.642012 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-s2r7z" event={"ID":"5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9","Type":"ContainerStarted","Data":"f91bf6c32332b94e76ad7f44dae12d4e139d708c8446895e202898a80258a0ce"} Oct 13 09:13:50 crc kubenswrapper[4685]: I1013 09:13:50.657764 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-s2r7z" event={"ID":"5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9","Type":"ContainerStarted","Data":"a025fc36f30faa80ca0e606efd1154e4bdfc82486ddb3bca0997c26094ffc703"} Oct 13 09:13:50 crc kubenswrapper[4685]: I1013 09:13:50.724997 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-s2r7z" podStartSLOduration=2.185809976 podStartE2EDuration="2.724974176s" podCreationTimestamp="2025-10-13 09:13:48 +0000 UTC" firstStartedPulling="2025-10-13 09:13:49.532307455 +0000 UTC m=+1754.680183216" lastFinishedPulling="2025-10-13 09:13:50.071471654 +0000 UTC m=+1755.219347416" observedRunningTime="2025-10-13 09:13:50.685512072 +0000 UTC m=+1755.833387853" watchObservedRunningTime="2025-10-13 09:13:50.724974176 +0000 UTC m=+1755.872849937" Oct 13 09:13:55 crc kubenswrapper[4685]: I1013 09:13:55.510503 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:13:55 crc kubenswrapper[4685]: E1013 09:13:55.511585 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:13:55 crc kubenswrapper[4685]: I1013 09:13:55.564333 4685 scope.go:117] "RemoveContainer" containerID="94fa93c93a9ce7c794e14c1fbcc577fb3bd0ffa09823ea05fdf9f6ca44d188c4" Oct 13 09:13:55 crc kubenswrapper[4685]: I1013 09:13:55.600227 4685 scope.go:117] "RemoveContainer" containerID="e6a919343f2e0025221223526df6f1e6f23a7bf2d7bb5ecd786fe39aed46ab3c" Oct 13 09:13:55 crc kubenswrapper[4685]: I1013 09:13:55.644310 4685 scope.go:117] "RemoveContainer" containerID="a7eaea1b8683eb5def98f74abe4a0e16d74d1cc6b26849938ced531c5d1651a9" Oct 13 09:13:55 crc kubenswrapper[4685]: I1013 09:13:55.682183 4685 scope.go:117] "RemoveContainer" containerID="bcee7a4d5a46dd9d25c4ffcbe3993eecbcdb1a858588a54fcc9653da1624702b" Oct 13 09:13:55 crc kubenswrapper[4685]: I1013 09:13:55.746967 4685 scope.go:117] "RemoveContainer" containerID="c1b8da12450a58ae7a4ce71b7afc3b047747dc08dd9165b1c8a13d4c1f03ee0d" Oct 13 09:13:55 crc kubenswrapper[4685]: I1013 09:13:55.783479 4685 scope.go:117] "RemoveContainer" containerID="03354a870611f8b31e4d0c070c0cf8a0b4ec8f54af85a611dbc62a161518f252" Oct 13 09:13:59 crc kubenswrapper[4685]: I1013 09:13:59.503714 4685 scope.go:117] "RemoveContainer" containerID="f4bd2015aa86f0bf7c40324d231108b8eefc9288b86f37bc304379cab4fbde84" Oct 13 09:13:59 crc kubenswrapper[4685]: E1013 09:13:59.504237 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:14:06 crc kubenswrapper[4685]: I1013 09:14:06.502796 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:14:06 crc kubenswrapper[4685]: E1013 09:14:06.503894 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:14:08 crc kubenswrapper[4685]: I1013 09:14:08.041889 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-l4jzf"] Oct 13 09:14:08 crc kubenswrapper[4685]: I1013 09:14:08.058817 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-l4jzf"] Oct 13 09:14:09 crc kubenswrapper[4685]: I1013 09:14:09.516031 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11c738eb-e9be-4e64-a38c-da231145e8ab" path="/var/lib/kubelet/pods/11c738eb-e9be-4e64-a38c-da231145e8ab/volumes" Oct 13 09:14:14 crc kubenswrapper[4685]: I1013 09:14:14.503296 4685 scope.go:117] "RemoveContainer" containerID="f4bd2015aa86f0bf7c40324d231108b8eefc9288b86f37bc304379cab4fbde84" Oct 13 09:14:14 crc kubenswrapper[4685]: E1013 09:14:14.504240 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:14:18 crc kubenswrapper[4685]: I1013 09:14:18.503298 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:14:18 crc kubenswrapper[4685]: E1013 09:14:18.503968 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:14:27 crc kubenswrapper[4685]: I1013 09:14:27.505148 4685 scope.go:117] "RemoveContainer" containerID="f4bd2015aa86f0bf7c40324d231108b8eefc9288b86f37bc304379cab4fbde84" Oct 13 09:14:27 crc kubenswrapper[4685]: E1013 09:14:27.505959 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:14:31 crc kubenswrapper[4685]: I1013 09:14:31.048655 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-5psdz"] Oct 13 09:14:31 crc kubenswrapper[4685]: I1013 09:14:31.055392 4685 generic.go:334] "Generic (PLEG): container finished" podID="5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9" containerID="a025fc36f30faa80ca0e606efd1154e4bdfc82486ddb3bca0997c26094ffc703" exitCode=0 Oct 13 09:14:31 crc kubenswrapper[4685]: I1013 09:14:31.055443 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-s2r7z" event={"ID":"5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9","Type":"ContainerDied","Data":"a025fc36f30faa80ca0e606efd1154e4bdfc82486ddb3bca0997c26094ffc703"} Oct 13 09:14:31 crc kubenswrapper[4685]: I1013 09:14:31.065637 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-5psdz"] Oct 13 09:14:31 crc kubenswrapper[4685]: I1013 09:14:31.504603 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:14:31 crc kubenswrapper[4685]: E1013 09:14:31.505040 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:14:31 crc kubenswrapper[4685]: I1013 09:14:31.521558 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c0a4dc6-640e-4aea-9865-aadb943fcc6a" path="/var/lib/kubelet/pods/5c0a4dc6-640e-4aea-9865-aadb943fcc6a/volumes" Oct 13 09:14:32 crc kubenswrapper[4685]: I1013 09:14:32.030286 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-wvrpr"] Oct 13 09:14:32 crc kubenswrapper[4685]: I1013 09:14:32.038008 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-wvrpr"] Oct 13 09:14:32 crc kubenswrapper[4685]: I1013 09:14:32.510474 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-s2r7z" Oct 13 09:14:32 crc kubenswrapper[4685]: I1013 09:14:32.589696 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skk58\" (UniqueName: \"kubernetes.io/projected/5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9-kube-api-access-skk58\") pod \"5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9\" (UID: \"5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9\") " Oct 13 09:14:32 crc kubenswrapper[4685]: I1013 09:14:32.589770 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9-inventory\") pod \"5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9\" (UID: \"5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9\") " Oct 13 09:14:32 crc kubenswrapper[4685]: I1013 09:14:32.589997 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9-ssh-key\") pod \"5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9\" (UID: \"5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9\") " Oct 13 09:14:32 crc kubenswrapper[4685]: I1013 09:14:32.629439 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9" (UID: "5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:14:32 crc kubenswrapper[4685]: I1013 09:14:32.629475 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9-kube-api-access-skk58" (OuterVolumeSpecName: "kube-api-access-skk58") pod "5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9" (UID: "5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9"). InnerVolumeSpecName "kube-api-access-skk58". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:14:32 crc kubenswrapper[4685]: I1013 09:14:32.638645 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9-inventory" (OuterVolumeSpecName: "inventory") pod "5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9" (UID: "5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:14:32 crc kubenswrapper[4685]: I1013 09:14:32.692788 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 09:14:32 crc kubenswrapper[4685]: I1013 09:14:32.692823 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skk58\" (UniqueName: \"kubernetes.io/projected/5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9-kube-api-access-skk58\") on node \"crc\" DevicePath \"\"" Oct 13 09:14:32 crc kubenswrapper[4685]: I1013 09:14:32.692835 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.074760 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-s2r7z" event={"ID":"5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9","Type":"ContainerDied","Data":"f91bf6c32332b94e76ad7f44dae12d4e139d708c8446895e202898a80258a0ce"} Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.074797 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f91bf6c32332b94e76ad7f44dae12d4e139d708c8446895e202898a80258a0ce" Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.074834 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-s2r7z" Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.165757 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jhmms"] Oct 13 09:14:33 crc kubenswrapper[4685]: E1013 09:14:33.166114 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.166132 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.166327 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.166891 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jhmms" Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.170008 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.170444 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.170695 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hcn4z" Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.171039 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.177493 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jhmms"] Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.304088 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j29mf\" (UniqueName: \"kubernetes.io/projected/2a07788a-cfaa-4c89-91ec-96999f31a6fd-kube-api-access-j29mf\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jhmms\" (UID: \"2a07788a-cfaa-4c89-91ec-96999f31a6fd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jhmms" Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.304652 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a07788a-cfaa-4c89-91ec-96999f31a6fd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jhmms\" (UID: \"2a07788a-cfaa-4c89-91ec-96999f31a6fd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jhmms" Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.304824 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a07788a-cfaa-4c89-91ec-96999f31a6fd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jhmms\" (UID: \"2a07788a-cfaa-4c89-91ec-96999f31a6fd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jhmms" Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.406596 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j29mf\" (UniqueName: \"kubernetes.io/projected/2a07788a-cfaa-4c89-91ec-96999f31a6fd-kube-api-access-j29mf\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jhmms\" (UID: \"2a07788a-cfaa-4c89-91ec-96999f31a6fd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jhmms" Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.406687 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a07788a-cfaa-4c89-91ec-96999f31a6fd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jhmms\" (UID: \"2a07788a-cfaa-4c89-91ec-96999f31a6fd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jhmms" Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.406726 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a07788a-cfaa-4c89-91ec-96999f31a6fd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jhmms\" (UID: \"2a07788a-cfaa-4c89-91ec-96999f31a6fd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jhmms" Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.411267 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a07788a-cfaa-4c89-91ec-96999f31a6fd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jhmms\" (UID: \"2a07788a-cfaa-4c89-91ec-96999f31a6fd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jhmms" Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.411532 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a07788a-cfaa-4c89-91ec-96999f31a6fd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jhmms\" (UID: \"2a07788a-cfaa-4c89-91ec-96999f31a6fd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jhmms" Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.433415 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j29mf\" (UniqueName: \"kubernetes.io/projected/2a07788a-cfaa-4c89-91ec-96999f31a6fd-kube-api-access-j29mf\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jhmms\" (UID: \"2a07788a-cfaa-4c89-91ec-96999f31a6fd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jhmms" Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.501309 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jhmms" Oct 13 09:14:33 crc kubenswrapper[4685]: I1013 09:14:33.519555 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8d4a69c-f774-44b5-bb13-5340f07027be" path="/var/lib/kubelet/pods/b8d4a69c-f774-44b5-bb13-5340f07027be/volumes" Oct 13 09:14:34 crc kubenswrapper[4685]: I1013 09:14:34.022614 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jhmms"] Oct 13 09:14:34 crc kubenswrapper[4685]: I1013 09:14:34.082636 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jhmms" event={"ID":"2a07788a-cfaa-4c89-91ec-96999f31a6fd","Type":"ContainerStarted","Data":"4bf7146d28cf252fc15054aa059b7c5c0b1c44327c99027f00b8b9916c6684ae"} Oct 13 09:14:36 crc kubenswrapper[4685]: I1013 09:14:36.102680 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jhmms" event={"ID":"2a07788a-cfaa-4c89-91ec-96999f31a6fd","Type":"ContainerStarted","Data":"4c8a97dd2e7d57def52eda40573dd591e39210e5e5c2bafe44b41938dba2ceaa"} Oct 13 09:14:36 crc kubenswrapper[4685]: I1013 09:14:36.122975 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jhmms" podStartSLOduration=2.1948561189999998 podStartE2EDuration="3.122958317s" podCreationTimestamp="2025-10-13 09:14:33 +0000 UTC" firstStartedPulling="2025-10-13 09:14:34.027385144 +0000 UTC m=+1799.175260905" lastFinishedPulling="2025-10-13 09:14:34.955487332 +0000 UTC m=+1800.103363103" observedRunningTime="2025-10-13 09:14:36.122670009 +0000 UTC m=+1801.270545800" watchObservedRunningTime="2025-10-13 09:14:36.122958317 +0000 UTC m=+1801.270834078" Oct 13 09:14:41 crc kubenswrapper[4685]: I1013 09:14:41.506946 4685 scope.go:117] "RemoveContainer" containerID="f4bd2015aa86f0bf7c40324d231108b8eefc9288b86f37bc304379cab4fbde84" Oct 13 09:14:42 crc kubenswrapper[4685]: I1013 09:14:42.160408 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" event={"ID":"b001d17a-1aea-44ba-86c5-ba6b312156c1","Type":"ContainerStarted","Data":"4e57efd2a5e99def1a73daa5f8404b15401378bfed57465d070b35fe0a552d79"} Oct 13 09:14:42 crc kubenswrapper[4685]: I1013 09:14:42.160846 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:14:42 crc kubenswrapper[4685]: I1013 09:14:42.503880 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:14:42 crc kubenswrapper[4685]: E1013 09:14:42.504708 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:14:53 crc kubenswrapper[4685]: I1013 09:14:53.445747 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:14:53 crc kubenswrapper[4685]: I1013 09:14:53.505243 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:14:53 crc kubenswrapper[4685]: E1013 09:14:53.505811 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:14:55 crc kubenswrapper[4685]: I1013 09:14:55.910807 4685 scope.go:117] "RemoveContainer" containerID="a110131aaa1c74e0bcd8e70b66e2ec1b6e514ab7428b13aeff24b729d18c411c" Oct 13 09:14:55 crc kubenswrapper[4685]: I1013 09:14:55.966354 4685 scope.go:117] "RemoveContainer" containerID="cbef8ef43372806b6c937531fa6dc0625fe37f464ed56e90a130285aadb2a357" Oct 13 09:14:56 crc kubenswrapper[4685]: I1013 09:14:56.042788 4685 scope.go:117] "RemoveContainer" containerID="d4b0d7d6ad8eb44f0d4a7899d571ce12a286694326a5c8a81ecfa4b0231cabdb" Oct 13 09:15:00 crc kubenswrapper[4685]: I1013 09:15:00.154058 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339115-vc2t6"] Oct 13 09:15:00 crc kubenswrapper[4685]: I1013 09:15:00.156212 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339115-vc2t6" Oct 13 09:15:00 crc kubenswrapper[4685]: I1013 09:15:00.158515 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 13 09:15:00 crc kubenswrapper[4685]: I1013 09:15:00.158829 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 13 09:15:00 crc kubenswrapper[4685]: I1013 09:15:00.164033 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339115-vc2t6"] Oct 13 09:15:00 crc kubenswrapper[4685]: I1013 09:15:00.190085 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcnq7\" (UniqueName: \"kubernetes.io/projected/d087604f-a3e3-4da2-940d-b8e6c59de3bc-kube-api-access-xcnq7\") pod \"collect-profiles-29339115-vc2t6\" (UID: \"d087604f-a3e3-4da2-940d-b8e6c59de3bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339115-vc2t6" Oct 13 09:15:00 crc kubenswrapper[4685]: I1013 09:15:00.190252 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d087604f-a3e3-4da2-940d-b8e6c59de3bc-config-volume\") pod \"collect-profiles-29339115-vc2t6\" (UID: \"d087604f-a3e3-4da2-940d-b8e6c59de3bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339115-vc2t6" Oct 13 09:15:00 crc kubenswrapper[4685]: I1013 09:15:00.190404 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d087604f-a3e3-4da2-940d-b8e6c59de3bc-secret-volume\") pod \"collect-profiles-29339115-vc2t6\" (UID: \"d087604f-a3e3-4da2-940d-b8e6c59de3bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339115-vc2t6" Oct 13 09:15:00 crc kubenswrapper[4685]: I1013 09:15:00.293111 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcnq7\" (UniqueName: \"kubernetes.io/projected/d087604f-a3e3-4da2-940d-b8e6c59de3bc-kube-api-access-xcnq7\") pod \"collect-profiles-29339115-vc2t6\" (UID: \"d087604f-a3e3-4da2-940d-b8e6c59de3bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339115-vc2t6" Oct 13 09:15:00 crc kubenswrapper[4685]: I1013 09:15:00.293188 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d087604f-a3e3-4da2-940d-b8e6c59de3bc-config-volume\") pod \"collect-profiles-29339115-vc2t6\" (UID: \"d087604f-a3e3-4da2-940d-b8e6c59de3bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339115-vc2t6" Oct 13 09:15:00 crc kubenswrapper[4685]: I1013 09:15:00.293241 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d087604f-a3e3-4da2-940d-b8e6c59de3bc-secret-volume\") pod \"collect-profiles-29339115-vc2t6\" (UID: \"d087604f-a3e3-4da2-940d-b8e6c59de3bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339115-vc2t6" Oct 13 09:15:00 crc kubenswrapper[4685]: I1013 09:15:00.294341 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d087604f-a3e3-4da2-940d-b8e6c59de3bc-config-volume\") pod \"collect-profiles-29339115-vc2t6\" (UID: \"d087604f-a3e3-4da2-940d-b8e6c59de3bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339115-vc2t6" Oct 13 09:15:00 crc kubenswrapper[4685]: I1013 09:15:00.304760 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d087604f-a3e3-4da2-940d-b8e6c59de3bc-secret-volume\") pod \"collect-profiles-29339115-vc2t6\" (UID: \"d087604f-a3e3-4da2-940d-b8e6c59de3bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339115-vc2t6" Oct 13 09:15:00 crc kubenswrapper[4685]: I1013 09:15:00.309872 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcnq7\" (UniqueName: \"kubernetes.io/projected/d087604f-a3e3-4da2-940d-b8e6c59de3bc-kube-api-access-xcnq7\") pod \"collect-profiles-29339115-vc2t6\" (UID: \"d087604f-a3e3-4da2-940d-b8e6c59de3bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339115-vc2t6" Oct 13 09:15:00 crc kubenswrapper[4685]: I1013 09:15:00.483061 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339115-vc2t6" Oct 13 09:15:00 crc kubenswrapper[4685]: W1013 09:15:00.933540 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd087604f_a3e3_4da2_940d_b8e6c59de3bc.slice/crio-80267e27c5da740761491972048d41ede0b1f723dabaa171c85e92f69566d72b WatchSource:0}: Error finding container 80267e27c5da740761491972048d41ede0b1f723dabaa171c85e92f69566d72b: Status 404 returned error can't find the container with id 80267e27c5da740761491972048d41ede0b1f723dabaa171c85e92f69566d72b Oct 13 09:15:00 crc kubenswrapper[4685]: I1013 09:15:00.942672 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339115-vc2t6"] Oct 13 09:15:01 crc kubenswrapper[4685]: I1013 09:15:01.342885 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339115-vc2t6" event={"ID":"d087604f-a3e3-4da2-940d-b8e6c59de3bc","Type":"ContainerStarted","Data":"6a5a9ee2a4718abb7932d54c5db4e57de017d62616a663511ef401a547576c31"} Oct 13 09:15:01 crc kubenswrapper[4685]: I1013 09:15:01.343314 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339115-vc2t6" event={"ID":"d087604f-a3e3-4da2-940d-b8e6c59de3bc","Type":"ContainerStarted","Data":"80267e27c5da740761491972048d41ede0b1f723dabaa171c85e92f69566d72b"} Oct 13 09:15:01 crc kubenswrapper[4685]: I1013 09:15:01.361272 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29339115-vc2t6" podStartSLOduration=1.361250175 podStartE2EDuration="1.361250175s" podCreationTimestamp="2025-10-13 09:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:15:01.356706151 +0000 UTC m=+1826.504581952" watchObservedRunningTime="2025-10-13 09:15:01.361250175 +0000 UTC m=+1826.509125946" Oct 13 09:15:02 crc kubenswrapper[4685]: I1013 09:15:02.357698 4685 generic.go:334] "Generic (PLEG): container finished" podID="d087604f-a3e3-4da2-940d-b8e6c59de3bc" containerID="6a5a9ee2a4718abb7932d54c5db4e57de017d62616a663511ef401a547576c31" exitCode=0 Oct 13 09:15:02 crc kubenswrapper[4685]: I1013 09:15:02.357771 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339115-vc2t6" event={"ID":"d087604f-a3e3-4da2-940d-b8e6c59de3bc","Type":"ContainerDied","Data":"6a5a9ee2a4718abb7932d54c5db4e57de017d62616a663511ef401a547576c31"} Oct 13 09:15:03 crc kubenswrapper[4685]: I1013 09:15:03.719774 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339115-vc2t6" Oct 13 09:15:03 crc kubenswrapper[4685]: I1013 09:15:03.757226 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcnq7\" (UniqueName: \"kubernetes.io/projected/d087604f-a3e3-4da2-940d-b8e6c59de3bc-kube-api-access-xcnq7\") pod \"d087604f-a3e3-4da2-940d-b8e6c59de3bc\" (UID: \"d087604f-a3e3-4da2-940d-b8e6c59de3bc\") " Oct 13 09:15:03 crc kubenswrapper[4685]: I1013 09:15:03.757744 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d087604f-a3e3-4da2-940d-b8e6c59de3bc-secret-volume\") pod \"d087604f-a3e3-4da2-940d-b8e6c59de3bc\" (UID: \"d087604f-a3e3-4da2-940d-b8e6c59de3bc\") " Oct 13 09:15:03 crc kubenswrapper[4685]: I1013 09:15:03.757825 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d087604f-a3e3-4da2-940d-b8e6c59de3bc-config-volume\") pod \"d087604f-a3e3-4da2-940d-b8e6c59de3bc\" (UID: \"d087604f-a3e3-4da2-940d-b8e6c59de3bc\") " Oct 13 09:15:03 crc kubenswrapper[4685]: I1013 09:15:03.758528 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d087604f-a3e3-4da2-940d-b8e6c59de3bc-config-volume" (OuterVolumeSpecName: "config-volume") pod "d087604f-a3e3-4da2-940d-b8e6c59de3bc" (UID: "d087604f-a3e3-4da2-940d-b8e6c59de3bc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:15:03 crc kubenswrapper[4685]: I1013 09:15:03.764779 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d087604f-a3e3-4da2-940d-b8e6c59de3bc-kube-api-access-xcnq7" (OuterVolumeSpecName: "kube-api-access-xcnq7") pod "d087604f-a3e3-4da2-940d-b8e6c59de3bc" (UID: "d087604f-a3e3-4da2-940d-b8e6c59de3bc"). InnerVolumeSpecName "kube-api-access-xcnq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:15:03 crc kubenswrapper[4685]: I1013 09:15:03.764833 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d087604f-a3e3-4da2-940d-b8e6c59de3bc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d087604f-a3e3-4da2-940d-b8e6c59de3bc" (UID: "d087604f-a3e3-4da2-940d-b8e6c59de3bc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:15:03 crc kubenswrapper[4685]: I1013 09:15:03.859981 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcnq7\" (UniqueName: \"kubernetes.io/projected/d087604f-a3e3-4da2-940d-b8e6c59de3bc-kube-api-access-xcnq7\") on node \"crc\" DevicePath \"\"" Oct 13 09:15:03 crc kubenswrapper[4685]: I1013 09:15:03.860032 4685 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d087604f-a3e3-4da2-940d-b8e6c59de3bc-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 13 09:15:03 crc kubenswrapper[4685]: I1013 09:15:03.860049 4685 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d087604f-a3e3-4da2-940d-b8e6c59de3bc-config-volume\") on node \"crc\" DevicePath \"\"" Oct 13 09:15:04 crc kubenswrapper[4685]: I1013 09:15:04.376856 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339115-vc2t6" event={"ID":"d087604f-a3e3-4da2-940d-b8e6c59de3bc","Type":"ContainerDied","Data":"80267e27c5da740761491972048d41ede0b1f723dabaa171c85e92f69566d72b"} Oct 13 09:15:04 crc kubenswrapper[4685]: I1013 09:15:04.376909 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80267e27c5da740761491972048d41ede0b1f723dabaa171c85e92f69566d72b" Oct 13 09:15:04 crc kubenswrapper[4685]: I1013 09:15:04.377014 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339115-vc2t6" Oct 13 09:15:04 crc kubenswrapper[4685]: I1013 09:15:04.504066 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:15:04 crc kubenswrapper[4685]: E1013 09:15:04.504377 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:15:15 crc kubenswrapper[4685]: I1013 09:15:15.513470 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:15:15 crc kubenswrapper[4685]: E1013 09:15:15.514356 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:15:18 crc kubenswrapper[4685]: I1013 09:15:18.063219 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-tkkqw"] Oct 13 09:15:18 crc kubenswrapper[4685]: I1013 09:15:18.074955 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-tkkqw"] Oct 13 09:15:19 crc kubenswrapper[4685]: I1013 09:15:19.517020 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fae68cb4-7110-40c0-b48b-b38036250c61" path="/var/lib/kubelet/pods/fae68cb4-7110-40c0-b48b-b38036250c61/volumes" Oct 13 09:15:29 crc kubenswrapper[4685]: I1013 09:15:29.504000 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:15:29 crc kubenswrapper[4685]: E1013 09:15:29.504636 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:15:31 crc kubenswrapper[4685]: I1013 09:15:31.647147 4685 generic.go:334] "Generic (PLEG): container finished" podID="2a07788a-cfaa-4c89-91ec-96999f31a6fd" containerID="4c8a97dd2e7d57def52eda40573dd591e39210e5e5c2bafe44b41938dba2ceaa" exitCode=2 Oct 13 09:15:31 crc kubenswrapper[4685]: I1013 09:15:31.647236 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jhmms" event={"ID":"2a07788a-cfaa-4c89-91ec-96999f31a6fd","Type":"ContainerDied","Data":"4c8a97dd2e7d57def52eda40573dd591e39210e5e5c2bafe44b41938dba2ceaa"} Oct 13 09:15:33 crc kubenswrapper[4685]: I1013 09:15:33.076335 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jhmms" Oct 13 09:15:33 crc kubenswrapper[4685]: I1013 09:15:33.146238 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a07788a-cfaa-4c89-91ec-96999f31a6fd-inventory\") pod \"2a07788a-cfaa-4c89-91ec-96999f31a6fd\" (UID: \"2a07788a-cfaa-4c89-91ec-96999f31a6fd\") " Oct 13 09:15:33 crc kubenswrapper[4685]: I1013 09:15:33.146333 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j29mf\" (UniqueName: \"kubernetes.io/projected/2a07788a-cfaa-4c89-91ec-96999f31a6fd-kube-api-access-j29mf\") pod \"2a07788a-cfaa-4c89-91ec-96999f31a6fd\" (UID: \"2a07788a-cfaa-4c89-91ec-96999f31a6fd\") " Oct 13 09:15:33 crc kubenswrapper[4685]: I1013 09:15:33.146421 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a07788a-cfaa-4c89-91ec-96999f31a6fd-ssh-key\") pod \"2a07788a-cfaa-4c89-91ec-96999f31a6fd\" (UID: \"2a07788a-cfaa-4c89-91ec-96999f31a6fd\") " Oct 13 09:15:33 crc kubenswrapper[4685]: I1013 09:15:33.155148 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a07788a-cfaa-4c89-91ec-96999f31a6fd-kube-api-access-j29mf" (OuterVolumeSpecName: "kube-api-access-j29mf") pod "2a07788a-cfaa-4c89-91ec-96999f31a6fd" (UID: "2a07788a-cfaa-4c89-91ec-96999f31a6fd"). InnerVolumeSpecName "kube-api-access-j29mf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:15:33 crc kubenswrapper[4685]: I1013 09:15:33.178591 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a07788a-cfaa-4c89-91ec-96999f31a6fd-inventory" (OuterVolumeSpecName: "inventory") pod "2a07788a-cfaa-4c89-91ec-96999f31a6fd" (UID: "2a07788a-cfaa-4c89-91ec-96999f31a6fd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:15:33 crc kubenswrapper[4685]: I1013 09:15:33.189836 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a07788a-cfaa-4c89-91ec-96999f31a6fd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2a07788a-cfaa-4c89-91ec-96999f31a6fd" (UID: "2a07788a-cfaa-4c89-91ec-96999f31a6fd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:15:33 crc kubenswrapper[4685]: I1013 09:15:33.247964 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j29mf\" (UniqueName: \"kubernetes.io/projected/2a07788a-cfaa-4c89-91ec-96999f31a6fd-kube-api-access-j29mf\") on node \"crc\" DevicePath \"\"" Oct 13 09:15:33 crc kubenswrapper[4685]: I1013 09:15:33.248001 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a07788a-cfaa-4c89-91ec-96999f31a6fd-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 09:15:33 crc kubenswrapper[4685]: I1013 09:15:33.248015 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a07788a-cfaa-4c89-91ec-96999f31a6fd-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 09:15:33 crc kubenswrapper[4685]: I1013 09:15:33.667340 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jhmms" event={"ID":"2a07788a-cfaa-4c89-91ec-96999f31a6fd","Type":"ContainerDied","Data":"4bf7146d28cf252fc15054aa059b7c5c0b1c44327c99027f00b8b9916c6684ae"} Oct 13 09:15:33 crc kubenswrapper[4685]: I1013 09:15:33.667373 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4bf7146d28cf252fc15054aa059b7c5c0b1c44327c99027f00b8b9916c6684ae" Oct 13 09:15:33 crc kubenswrapper[4685]: I1013 09:15:33.667499 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jhmms" Oct 13 09:15:40 crc kubenswrapper[4685]: I1013 09:15:40.029728 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jk24b"] Oct 13 09:15:40 crc kubenswrapper[4685]: E1013 09:15:40.030704 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a07788a-cfaa-4c89-91ec-96999f31a6fd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 13 09:15:40 crc kubenswrapper[4685]: I1013 09:15:40.030722 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a07788a-cfaa-4c89-91ec-96999f31a6fd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 13 09:15:40 crc kubenswrapper[4685]: E1013 09:15:40.030736 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d087604f-a3e3-4da2-940d-b8e6c59de3bc" containerName="collect-profiles" Oct 13 09:15:40 crc kubenswrapper[4685]: I1013 09:15:40.030742 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="d087604f-a3e3-4da2-940d-b8e6c59de3bc" containerName="collect-profiles" Oct 13 09:15:40 crc kubenswrapper[4685]: I1013 09:15:40.030900 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="d087604f-a3e3-4da2-940d-b8e6c59de3bc" containerName="collect-profiles" Oct 13 09:15:40 crc kubenswrapper[4685]: I1013 09:15:40.030932 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a07788a-cfaa-4c89-91ec-96999f31a6fd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 13 09:15:40 crc kubenswrapper[4685]: I1013 09:15:40.031658 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jk24b" Oct 13 09:15:40 crc kubenswrapper[4685]: I1013 09:15:40.034044 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hcn4z" Oct 13 09:15:40 crc kubenswrapper[4685]: I1013 09:15:40.034285 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 09:15:40 crc kubenswrapper[4685]: I1013 09:15:40.034627 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 09:15:40 crc kubenswrapper[4685]: I1013 09:15:40.034897 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 09:15:40 crc kubenswrapper[4685]: I1013 09:15:40.060445 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jk24b"] Oct 13 09:15:40 crc kubenswrapper[4685]: I1013 09:15:40.199212 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdc6l\" (UniqueName: \"kubernetes.io/projected/c9f2aa20-f6e6-47e7-9244-36184ce2cf1c-kube-api-access-zdc6l\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jk24b\" (UID: \"c9f2aa20-f6e6-47e7-9244-36184ce2cf1c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jk24b" Oct 13 09:15:40 crc kubenswrapper[4685]: I1013 09:15:40.199499 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9f2aa20-f6e6-47e7-9244-36184ce2cf1c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jk24b\" (UID: \"c9f2aa20-f6e6-47e7-9244-36184ce2cf1c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jk24b" Oct 13 09:15:40 crc kubenswrapper[4685]: I1013 09:15:40.199548 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9f2aa20-f6e6-47e7-9244-36184ce2cf1c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jk24b\" (UID: \"c9f2aa20-f6e6-47e7-9244-36184ce2cf1c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jk24b" Oct 13 09:15:40 crc kubenswrapper[4685]: I1013 09:15:40.301932 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdc6l\" (UniqueName: \"kubernetes.io/projected/c9f2aa20-f6e6-47e7-9244-36184ce2cf1c-kube-api-access-zdc6l\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jk24b\" (UID: \"c9f2aa20-f6e6-47e7-9244-36184ce2cf1c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jk24b" Oct 13 09:15:40 crc kubenswrapper[4685]: I1013 09:15:40.302093 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9f2aa20-f6e6-47e7-9244-36184ce2cf1c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jk24b\" (UID: \"c9f2aa20-f6e6-47e7-9244-36184ce2cf1c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jk24b" Oct 13 09:15:40 crc kubenswrapper[4685]: I1013 09:15:40.302125 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9f2aa20-f6e6-47e7-9244-36184ce2cf1c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jk24b\" (UID: \"c9f2aa20-f6e6-47e7-9244-36184ce2cf1c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jk24b" Oct 13 09:15:40 crc kubenswrapper[4685]: I1013 09:15:40.309013 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9f2aa20-f6e6-47e7-9244-36184ce2cf1c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jk24b\" (UID: \"c9f2aa20-f6e6-47e7-9244-36184ce2cf1c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jk24b" Oct 13 09:15:40 crc kubenswrapper[4685]: I1013 09:15:40.312709 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9f2aa20-f6e6-47e7-9244-36184ce2cf1c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jk24b\" (UID: \"c9f2aa20-f6e6-47e7-9244-36184ce2cf1c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jk24b" Oct 13 09:15:40 crc kubenswrapper[4685]: I1013 09:15:40.332499 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdc6l\" (UniqueName: \"kubernetes.io/projected/c9f2aa20-f6e6-47e7-9244-36184ce2cf1c-kube-api-access-zdc6l\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jk24b\" (UID: \"c9f2aa20-f6e6-47e7-9244-36184ce2cf1c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jk24b" Oct 13 09:15:40 crc kubenswrapper[4685]: I1013 09:15:40.350089 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jk24b" Oct 13 09:15:40 crc kubenswrapper[4685]: I1013 09:15:40.943798 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jk24b"] Oct 13 09:15:41 crc kubenswrapper[4685]: I1013 09:15:41.751491 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jk24b" event={"ID":"c9f2aa20-f6e6-47e7-9244-36184ce2cf1c","Type":"ContainerStarted","Data":"4c049a118ef74cb57c489d01f058e77e15f1494e5259546bffef3d4d0327c837"} Oct 13 09:15:41 crc kubenswrapper[4685]: I1013 09:15:41.752116 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jk24b" event={"ID":"c9f2aa20-f6e6-47e7-9244-36184ce2cf1c","Type":"ContainerStarted","Data":"e7802715d3d195487033225fd69e382d41faeace24c57e397eb2b3bc04e15253"} Oct 13 09:15:41 crc kubenswrapper[4685]: I1013 09:15:41.767742 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jk24b" podStartSLOduration=1.2002993530000001 podStartE2EDuration="1.767723942s" podCreationTimestamp="2025-10-13 09:15:40 +0000 UTC" firstStartedPulling="2025-10-13 09:15:40.939942956 +0000 UTC m=+1866.087818717" lastFinishedPulling="2025-10-13 09:15:41.507367525 +0000 UTC m=+1866.655243306" observedRunningTime="2025-10-13 09:15:41.766831888 +0000 UTC m=+1866.914707649" watchObservedRunningTime="2025-10-13 09:15:41.767723942 +0000 UTC m=+1866.915599703" Oct 13 09:15:44 crc kubenswrapper[4685]: I1013 09:15:44.504298 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:15:44 crc kubenswrapper[4685]: E1013 09:15:44.504858 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:15:56 crc kubenswrapper[4685]: I1013 09:15:56.176219 4685 scope.go:117] "RemoveContainer" containerID="4387f4f578b34308748b28e4f9426909a73eca7b00aaa8820ff473823a37df64" Oct 13 09:15:59 crc kubenswrapper[4685]: I1013 09:15:59.503311 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:15:59 crc kubenswrapper[4685]: E1013 09:15:59.503740 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:16:10 crc kubenswrapper[4685]: I1013 09:16:10.502588 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:16:10 crc kubenswrapper[4685]: E1013 09:16:10.503372 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:16:24 crc kubenswrapper[4685]: I1013 09:16:24.502770 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:16:24 crc kubenswrapper[4685]: E1013 09:16:24.503510 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:16:31 crc kubenswrapper[4685]: I1013 09:16:31.206873 4685 generic.go:334] "Generic (PLEG): container finished" podID="c9f2aa20-f6e6-47e7-9244-36184ce2cf1c" containerID="4c049a118ef74cb57c489d01f058e77e15f1494e5259546bffef3d4d0327c837" exitCode=0 Oct 13 09:16:31 crc kubenswrapper[4685]: I1013 09:16:31.207016 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jk24b" event={"ID":"c9f2aa20-f6e6-47e7-9244-36184ce2cf1c","Type":"ContainerDied","Data":"4c049a118ef74cb57c489d01f058e77e15f1494e5259546bffef3d4d0327c837"} Oct 13 09:16:32 crc kubenswrapper[4685]: I1013 09:16:32.640270 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jk24b" Oct 13 09:16:32 crc kubenswrapper[4685]: I1013 09:16:32.749163 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdc6l\" (UniqueName: \"kubernetes.io/projected/c9f2aa20-f6e6-47e7-9244-36184ce2cf1c-kube-api-access-zdc6l\") pod \"c9f2aa20-f6e6-47e7-9244-36184ce2cf1c\" (UID: \"c9f2aa20-f6e6-47e7-9244-36184ce2cf1c\") " Oct 13 09:16:32 crc kubenswrapper[4685]: I1013 09:16:32.749375 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9f2aa20-f6e6-47e7-9244-36184ce2cf1c-inventory\") pod \"c9f2aa20-f6e6-47e7-9244-36184ce2cf1c\" (UID: \"c9f2aa20-f6e6-47e7-9244-36184ce2cf1c\") " Oct 13 09:16:32 crc kubenswrapper[4685]: I1013 09:16:32.749440 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9f2aa20-f6e6-47e7-9244-36184ce2cf1c-ssh-key\") pod \"c9f2aa20-f6e6-47e7-9244-36184ce2cf1c\" (UID: \"c9f2aa20-f6e6-47e7-9244-36184ce2cf1c\") " Oct 13 09:16:32 crc kubenswrapper[4685]: I1013 09:16:32.755251 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9f2aa20-f6e6-47e7-9244-36184ce2cf1c-kube-api-access-zdc6l" (OuterVolumeSpecName: "kube-api-access-zdc6l") pod "c9f2aa20-f6e6-47e7-9244-36184ce2cf1c" (UID: "c9f2aa20-f6e6-47e7-9244-36184ce2cf1c"). InnerVolumeSpecName "kube-api-access-zdc6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:16:32 crc kubenswrapper[4685]: I1013 09:16:32.782675 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9f2aa20-f6e6-47e7-9244-36184ce2cf1c-inventory" (OuterVolumeSpecName: "inventory") pod "c9f2aa20-f6e6-47e7-9244-36184ce2cf1c" (UID: "c9f2aa20-f6e6-47e7-9244-36184ce2cf1c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:16:32 crc kubenswrapper[4685]: I1013 09:16:32.782691 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9f2aa20-f6e6-47e7-9244-36184ce2cf1c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c9f2aa20-f6e6-47e7-9244-36184ce2cf1c" (UID: "c9f2aa20-f6e6-47e7-9244-36184ce2cf1c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:16:32 crc kubenswrapper[4685]: I1013 09:16:32.851306 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9f2aa20-f6e6-47e7-9244-36184ce2cf1c-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 09:16:32 crc kubenswrapper[4685]: I1013 09:16:32.851347 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9f2aa20-f6e6-47e7-9244-36184ce2cf1c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 09:16:32 crc kubenswrapper[4685]: I1013 09:16:32.851362 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdc6l\" (UniqueName: \"kubernetes.io/projected/c9f2aa20-f6e6-47e7-9244-36184ce2cf1c-kube-api-access-zdc6l\") on node \"crc\" DevicePath \"\"" Oct 13 09:16:33 crc kubenswrapper[4685]: I1013 09:16:33.226214 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jk24b" event={"ID":"c9f2aa20-f6e6-47e7-9244-36184ce2cf1c","Type":"ContainerDied","Data":"e7802715d3d195487033225fd69e382d41faeace24c57e397eb2b3bc04e15253"} Oct 13 09:16:33 crc kubenswrapper[4685]: I1013 09:16:33.226516 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7802715d3d195487033225fd69e382d41faeace24c57e397eb2b3bc04e15253" Oct 13 09:16:33 crc kubenswrapper[4685]: I1013 09:16:33.226325 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jk24b" Oct 13 09:16:33 crc kubenswrapper[4685]: I1013 09:16:33.333212 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-xsmbj"] Oct 13 09:16:33 crc kubenswrapper[4685]: E1013 09:16:33.333710 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9f2aa20-f6e6-47e7-9244-36184ce2cf1c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 13 09:16:33 crc kubenswrapper[4685]: I1013 09:16:33.333731 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9f2aa20-f6e6-47e7-9244-36184ce2cf1c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 13 09:16:33 crc kubenswrapper[4685]: I1013 09:16:33.333990 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9f2aa20-f6e6-47e7-9244-36184ce2cf1c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 13 09:16:33 crc kubenswrapper[4685]: I1013 09:16:33.334849 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-xsmbj" Oct 13 09:16:33 crc kubenswrapper[4685]: I1013 09:16:33.337415 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 09:16:33 crc kubenswrapper[4685]: I1013 09:16:33.337652 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 09:16:33 crc kubenswrapper[4685]: I1013 09:16:33.339460 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hcn4z" Oct 13 09:16:33 crc kubenswrapper[4685]: I1013 09:16:33.346637 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-xsmbj"] Oct 13 09:16:33 crc kubenswrapper[4685]: I1013 09:16:33.346883 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 09:16:33 crc kubenswrapper[4685]: I1013 09:16:33.462878 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/73114199-2f57-451c-8b19-81ad0fbb98f5-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-xsmbj\" (UID: \"73114199-2f57-451c-8b19-81ad0fbb98f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-xsmbj" Oct 13 09:16:33 crc kubenswrapper[4685]: I1013 09:16:33.463132 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/73114199-2f57-451c-8b19-81ad0fbb98f5-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-xsmbj\" (UID: \"73114199-2f57-451c-8b19-81ad0fbb98f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-xsmbj" Oct 13 09:16:33 crc kubenswrapper[4685]: I1013 09:16:33.463223 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4fvz\" (UniqueName: \"kubernetes.io/projected/73114199-2f57-451c-8b19-81ad0fbb98f5-kube-api-access-n4fvz\") pod \"ssh-known-hosts-edpm-deployment-xsmbj\" (UID: \"73114199-2f57-451c-8b19-81ad0fbb98f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-xsmbj" Oct 13 09:16:33 crc kubenswrapper[4685]: I1013 09:16:33.565376 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/73114199-2f57-451c-8b19-81ad0fbb98f5-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-xsmbj\" (UID: \"73114199-2f57-451c-8b19-81ad0fbb98f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-xsmbj" Oct 13 09:16:33 crc kubenswrapper[4685]: I1013 09:16:33.565581 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/73114199-2f57-451c-8b19-81ad0fbb98f5-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-xsmbj\" (UID: \"73114199-2f57-451c-8b19-81ad0fbb98f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-xsmbj" Oct 13 09:16:33 crc kubenswrapper[4685]: I1013 09:16:33.565610 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4fvz\" (UniqueName: \"kubernetes.io/projected/73114199-2f57-451c-8b19-81ad0fbb98f5-kube-api-access-n4fvz\") pod \"ssh-known-hosts-edpm-deployment-xsmbj\" (UID: \"73114199-2f57-451c-8b19-81ad0fbb98f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-xsmbj" Oct 13 09:16:33 crc kubenswrapper[4685]: I1013 09:16:33.575405 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/73114199-2f57-451c-8b19-81ad0fbb98f5-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-xsmbj\" (UID: \"73114199-2f57-451c-8b19-81ad0fbb98f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-xsmbj" Oct 13 09:16:33 crc kubenswrapper[4685]: I1013 09:16:33.580574 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/73114199-2f57-451c-8b19-81ad0fbb98f5-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-xsmbj\" (UID: \"73114199-2f57-451c-8b19-81ad0fbb98f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-xsmbj" Oct 13 09:16:33 crc kubenswrapper[4685]: I1013 09:16:33.590797 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4fvz\" (UniqueName: \"kubernetes.io/projected/73114199-2f57-451c-8b19-81ad0fbb98f5-kube-api-access-n4fvz\") pod \"ssh-known-hosts-edpm-deployment-xsmbj\" (UID: \"73114199-2f57-451c-8b19-81ad0fbb98f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-xsmbj" Oct 13 09:16:33 crc kubenswrapper[4685]: I1013 09:16:33.660544 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-xsmbj" Oct 13 09:16:34 crc kubenswrapper[4685]: I1013 09:16:34.206348 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-xsmbj"] Oct 13 09:16:34 crc kubenswrapper[4685]: I1013 09:16:34.219175 4685 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 13 09:16:34 crc kubenswrapper[4685]: I1013 09:16:34.237073 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-xsmbj" event={"ID":"73114199-2f57-451c-8b19-81ad0fbb98f5","Type":"ContainerStarted","Data":"e6914d6d6e62f32d60cff21bd06400652550ea73fdd27bc99db11b399398c40d"} Oct 13 09:16:35 crc kubenswrapper[4685]: I1013 09:16:35.245780 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-xsmbj" event={"ID":"73114199-2f57-451c-8b19-81ad0fbb98f5","Type":"ContainerStarted","Data":"49daa33b9677558a4f9155f053f2eafbe261b083a584e6036a97b56abaabfb9a"} Oct 13 09:16:35 crc kubenswrapper[4685]: I1013 09:16:35.270685 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-xsmbj" podStartSLOduration=1.841171751 podStartE2EDuration="2.27066842s" podCreationTimestamp="2025-10-13 09:16:33 +0000 UTC" firstStartedPulling="2025-10-13 09:16:34.218772147 +0000 UTC m=+1919.366647928" lastFinishedPulling="2025-10-13 09:16:34.648268836 +0000 UTC m=+1919.796144597" observedRunningTime="2025-10-13 09:16:35.267692009 +0000 UTC m=+1920.415567780" watchObservedRunningTime="2025-10-13 09:16:35.27066842 +0000 UTC m=+1920.418544181" Oct 13 09:16:35 crc kubenswrapper[4685]: I1013 09:16:35.508977 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:16:35 crc kubenswrapper[4685]: E1013 09:16:35.509468 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:16:42 crc kubenswrapper[4685]: I1013 09:16:42.336724 4685 generic.go:334] "Generic (PLEG): container finished" podID="73114199-2f57-451c-8b19-81ad0fbb98f5" containerID="49daa33b9677558a4f9155f053f2eafbe261b083a584e6036a97b56abaabfb9a" exitCode=0 Oct 13 09:16:42 crc kubenswrapper[4685]: I1013 09:16:42.336947 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-xsmbj" event={"ID":"73114199-2f57-451c-8b19-81ad0fbb98f5","Type":"ContainerDied","Data":"49daa33b9677558a4f9155f053f2eafbe261b083a584e6036a97b56abaabfb9a"} Oct 13 09:16:43 crc kubenswrapper[4685]: I1013 09:16:43.813685 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-xsmbj" Oct 13 09:16:43 crc kubenswrapper[4685]: I1013 09:16:43.896528 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4fvz\" (UniqueName: \"kubernetes.io/projected/73114199-2f57-451c-8b19-81ad0fbb98f5-kube-api-access-n4fvz\") pod \"73114199-2f57-451c-8b19-81ad0fbb98f5\" (UID: \"73114199-2f57-451c-8b19-81ad0fbb98f5\") " Oct 13 09:16:43 crc kubenswrapper[4685]: I1013 09:16:43.896615 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/73114199-2f57-451c-8b19-81ad0fbb98f5-inventory-0\") pod \"73114199-2f57-451c-8b19-81ad0fbb98f5\" (UID: \"73114199-2f57-451c-8b19-81ad0fbb98f5\") " Oct 13 09:16:43 crc kubenswrapper[4685]: I1013 09:16:43.896742 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/73114199-2f57-451c-8b19-81ad0fbb98f5-ssh-key-openstack-edpm-ipam\") pod \"73114199-2f57-451c-8b19-81ad0fbb98f5\" (UID: \"73114199-2f57-451c-8b19-81ad0fbb98f5\") " Oct 13 09:16:43 crc kubenswrapper[4685]: I1013 09:16:43.902538 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73114199-2f57-451c-8b19-81ad0fbb98f5-kube-api-access-n4fvz" (OuterVolumeSpecName: "kube-api-access-n4fvz") pod "73114199-2f57-451c-8b19-81ad0fbb98f5" (UID: "73114199-2f57-451c-8b19-81ad0fbb98f5"). InnerVolumeSpecName "kube-api-access-n4fvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:16:43 crc kubenswrapper[4685]: I1013 09:16:43.925372 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73114199-2f57-451c-8b19-81ad0fbb98f5-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "73114199-2f57-451c-8b19-81ad0fbb98f5" (UID: "73114199-2f57-451c-8b19-81ad0fbb98f5"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:16:43 crc kubenswrapper[4685]: I1013 09:16:43.929329 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73114199-2f57-451c-8b19-81ad0fbb98f5-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "73114199-2f57-451c-8b19-81ad0fbb98f5" (UID: "73114199-2f57-451c-8b19-81ad0fbb98f5"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:16:43 crc kubenswrapper[4685]: I1013 09:16:43.998701 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/73114199-2f57-451c-8b19-81ad0fbb98f5-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 13 09:16:43 crc kubenswrapper[4685]: I1013 09:16:43.998743 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4fvz\" (UniqueName: \"kubernetes.io/projected/73114199-2f57-451c-8b19-81ad0fbb98f5-kube-api-access-n4fvz\") on node \"crc\" DevicePath \"\"" Oct 13 09:16:43 crc kubenswrapper[4685]: I1013 09:16:43.998756 4685 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/73114199-2f57-451c-8b19-81ad0fbb98f5-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 13 09:16:44 crc kubenswrapper[4685]: I1013 09:16:44.358531 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-xsmbj" Oct 13 09:16:44 crc kubenswrapper[4685]: I1013 09:16:44.358567 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-xsmbj" event={"ID":"73114199-2f57-451c-8b19-81ad0fbb98f5","Type":"ContainerDied","Data":"e6914d6d6e62f32d60cff21bd06400652550ea73fdd27bc99db11b399398c40d"} Oct 13 09:16:44 crc kubenswrapper[4685]: I1013 09:16:44.358613 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6914d6d6e62f32d60cff21bd06400652550ea73fdd27bc99db11b399398c40d" Oct 13 09:16:44 crc kubenswrapper[4685]: I1013 09:16:44.438263 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-p2spf"] Oct 13 09:16:44 crc kubenswrapper[4685]: E1013 09:16:44.438707 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73114199-2f57-451c-8b19-81ad0fbb98f5" containerName="ssh-known-hosts-edpm-deployment" Oct 13 09:16:44 crc kubenswrapper[4685]: I1013 09:16:44.438733 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="73114199-2f57-451c-8b19-81ad0fbb98f5" containerName="ssh-known-hosts-edpm-deployment" Oct 13 09:16:44 crc kubenswrapper[4685]: I1013 09:16:44.438983 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="73114199-2f57-451c-8b19-81ad0fbb98f5" containerName="ssh-known-hosts-edpm-deployment" Oct 13 09:16:44 crc kubenswrapper[4685]: I1013 09:16:44.439571 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-p2spf" Oct 13 09:16:44 crc kubenswrapper[4685]: I1013 09:16:44.441587 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 09:16:44 crc kubenswrapper[4685]: I1013 09:16:44.441645 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 09:16:44 crc kubenswrapper[4685]: I1013 09:16:44.444891 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 09:16:44 crc kubenswrapper[4685]: I1013 09:16:44.452223 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-p2spf"] Oct 13 09:16:44 crc kubenswrapper[4685]: I1013 09:16:44.454594 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hcn4z" Oct 13 09:16:44 crc kubenswrapper[4685]: I1013 09:16:44.508608 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/82f29acf-d74b-434e-9aaf-7324cbc6c2d0-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-p2spf\" (UID: \"82f29acf-d74b-434e-9aaf-7324cbc6c2d0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-p2spf" Oct 13 09:16:44 crc kubenswrapper[4685]: I1013 09:16:44.508783 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6d5v\" (UniqueName: \"kubernetes.io/projected/82f29acf-d74b-434e-9aaf-7324cbc6c2d0-kube-api-access-j6d5v\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-p2spf\" (UID: \"82f29acf-d74b-434e-9aaf-7324cbc6c2d0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-p2spf" Oct 13 09:16:44 crc kubenswrapper[4685]: I1013 09:16:44.508812 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82f29acf-d74b-434e-9aaf-7324cbc6c2d0-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-p2spf\" (UID: \"82f29acf-d74b-434e-9aaf-7324cbc6c2d0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-p2spf" Oct 13 09:16:44 crc kubenswrapper[4685]: I1013 09:16:44.610289 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6d5v\" (UniqueName: \"kubernetes.io/projected/82f29acf-d74b-434e-9aaf-7324cbc6c2d0-kube-api-access-j6d5v\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-p2spf\" (UID: \"82f29acf-d74b-434e-9aaf-7324cbc6c2d0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-p2spf" Oct 13 09:16:44 crc kubenswrapper[4685]: I1013 09:16:44.610340 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82f29acf-d74b-434e-9aaf-7324cbc6c2d0-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-p2spf\" (UID: \"82f29acf-d74b-434e-9aaf-7324cbc6c2d0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-p2spf" Oct 13 09:16:44 crc kubenswrapper[4685]: I1013 09:16:44.610470 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/82f29acf-d74b-434e-9aaf-7324cbc6c2d0-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-p2spf\" (UID: \"82f29acf-d74b-434e-9aaf-7324cbc6c2d0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-p2spf" Oct 13 09:16:44 crc kubenswrapper[4685]: I1013 09:16:44.614769 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/82f29acf-d74b-434e-9aaf-7324cbc6c2d0-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-p2spf\" (UID: \"82f29acf-d74b-434e-9aaf-7324cbc6c2d0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-p2spf" Oct 13 09:16:44 crc kubenswrapper[4685]: I1013 09:16:44.622672 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82f29acf-d74b-434e-9aaf-7324cbc6c2d0-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-p2spf\" (UID: \"82f29acf-d74b-434e-9aaf-7324cbc6c2d0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-p2spf" Oct 13 09:16:44 crc kubenswrapper[4685]: I1013 09:16:44.631694 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6d5v\" (UniqueName: \"kubernetes.io/projected/82f29acf-d74b-434e-9aaf-7324cbc6c2d0-kube-api-access-j6d5v\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-p2spf\" (UID: \"82f29acf-d74b-434e-9aaf-7324cbc6c2d0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-p2spf" Oct 13 09:16:44 crc kubenswrapper[4685]: I1013 09:16:44.757780 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-p2spf" Oct 13 09:16:45 crc kubenswrapper[4685]: W1013 09:16:45.442775 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82f29acf_d74b_434e_9aaf_7324cbc6c2d0.slice/crio-170e8b896e2afe13967d5697f84706104fe2b5da5290c2c54e1f2782c7d13f4b WatchSource:0}: Error finding container 170e8b896e2afe13967d5697f84706104fe2b5da5290c2c54e1f2782c7d13f4b: Status 404 returned error can't find the container with id 170e8b896e2afe13967d5697f84706104fe2b5da5290c2c54e1f2782c7d13f4b Oct 13 09:16:45 crc kubenswrapper[4685]: I1013 09:16:45.455601 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-p2spf"] Oct 13 09:16:46 crc kubenswrapper[4685]: I1013 09:16:46.373003 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-p2spf" event={"ID":"82f29acf-d74b-434e-9aaf-7324cbc6c2d0","Type":"ContainerStarted","Data":"481a7bb9ebe0ba05ebef8a1a6ec6ee1d1e7a9dd968ed03bca56d03d6b53e42eb"} Oct 13 09:16:46 crc kubenswrapper[4685]: I1013 09:16:46.373230 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-p2spf" event={"ID":"82f29acf-d74b-434e-9aaf-7324cbc6c2d0","Type":"ContainerStarted","Data":"170e8b896e2afe13967d5697f84706104fe2b5da5290c2c54e1f2782c7d13f4b"} Oct 13 09:16:46 crc kubenswrapper[4685]: I1013 09:16:46.390086 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-p2spf" podStartSLOduration=1.909763147 podStartE2EDuration="2.39006872s" podCreationTimestamp="2025-10-13 09:16:44 +0000 UTC" firstStartedPulling="2025-10-13 09:16:45.446847387 +0000 UTC m=+1930.594723148" lastFinishedPulling="2025-10-13 09:16:45.92715296 +0000 UTC m=+1931.075028721" observedRunningTime="2025-10-13 09:16:46.388758364 +0000 UTC m=+1931.536634145" watchObservedRunningTime="2025-10-13 09:16:46.39006872 +0000 UTC m=+1931.537944491" Oct 13 09:16:49 crc kubenswrapper[4685]: I1013 09:16:49.502782 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:16:49 crc kubenswrapper[4685]: E1013 09:16:49.503414 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:16:55 crc kubenswrapper[4685]: I1013 09:16:55.456614 4685 generic.go:334] "Generic (PLEG): container finished" podID="82f29acf-d74b-434e-9aaf-7324cbc6c2d0" containerID="481a7bb9ebe0ba05ebef8a1a6ec6ee1d1e7a9dd968ed03bca56d03d6b53e42eb" exitCode=0 Oct 13 09:16:55 crc kubenswrapper[4685]: I1013 09:16:55.456662 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-p2spf" event={"ID":"82f29acf-d74b-434e-9aaf-7324cbc6c2d0","Type":"ContainerDied","Data":"481a7bb9ebe0ba05ebef8a1a6ec6ee1d1e7a9dd968ed03bca56d03d6b53e42eb"} Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.009365 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-p2spf" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.064851 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6d5v\" (UniqueName: \"kubernetes.io/projected/82f29acf-d74b-434e-9aaf-7324cbc6c2d0-kube-api-access-j6d5v\") pod \"82f29acf-d74b-434e-9aaf-7324cbc6c2d0\" (UID: \"82f29acf-d74b-434e-9aaf-7324cbc6c2d0\") " Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.065336 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/82f29acf-d74b-434e-9aaf-7324cbc6c2d0-ssh-key\") pod \"82f29acf-d74b-434e-9aaf-7324cbc6c2d0\" (UID: \"82f29acf-d74b-434e-9aaf-7324cbc6c2d0\") " Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.066101 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82f29acf-d74b-434e-9aaf-7324cbc6c2d0-inventory\") pod \"82f29acf-d74b-434e-9aaf-7324cbc6c2d0\" (UID: \"82f29acf-d74b-434e-9aaf-7324cbc6c2d0\") " Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.077116 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82f29acf-d74b-434e-9aaf-7324cbc6c2d0-kube-api-access-j6d5v" (OuterVolumeSpecName: "kube-api-access-j6d5v") pod "82f29acf-d74b-434e-9aaf-7324cbc6c2d0" (UID: "82f29acf-d74b-434e-9aaf-7324cbc6c2d0"). InnerVolumeSpecName "kube-api-access-j6d5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.093298 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82f29acf-d74b-434e-9aaf-7324cbc6c2d0-inventory" (OuterVolumeSpecName: "inventory") pod "82f29acf-d74b-434e-9aaf-7324cbc6c2d0" (UID: "82f29acf-d74b-434e-9aaf-7324cbc6c2d0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.116581 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82f29acf-d74b-434e-9aaf-7324cbc6c2d0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "82f29acf-d74b-434e-9aaf-7324cbc6c2d0" (UID: "82f29acf-d74b-434e-9aaf-7324cbc6c2d0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.168235 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82f29acf-d74b-434e-9aaf-7324cbc6c2d0-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.168266 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6d5v\" (UniqueName: \"kubernetes.io/projected/82f29acf-d74b-434e-9aaf-7324cbc6c2d0-kube-api-access-j6d5v\") on node \"crc\" DevicePath \"\"" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.168278 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/82f29acf-d74b-434e-9aaf-7324cbc6c2d0-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.479038 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-p2spf" event={"ID":"82f29acf-d74b-434e-9aaf-7324cbc6c2d0","Type":"ContainerDied","Data":"170e8b896e2afe13967d5697f84706104fe2b5da5290c2c54e1f2782c7d13f4b"} Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.479076 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="170e8b896e2afe13967d5697f84706104fe2b5da5290c2c54e1f2782c7d13f4b" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.479371 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-p2spf" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.561895 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj"] Oct 13 09:16:57 crc kubenswrapper[4685]: E1013 09:16:57.564140 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82f29acf-d74b-434e-9aaf-7324cbc6c2d0" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.564168 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="82f29acf-d74b-434e-9aaf-7324cbc6c2d0" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.564379 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="82f29acf-d74b-434e-9aaf-7324cbc6c2d0" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.565055 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.575391 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hcn4z" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.575645 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.575792 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.579754 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.585564 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj"] Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.678833 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6c7cd6b-3a10-490e-af5e-68a812d4b6f6-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj\" (UID: \"f6c7cd6b-3a10-490e-af5e-68a812d4b6f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.679135 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6c7cd6b-3a10-490e-af5e-68a812d4b6f6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj\" (UID: \"f6c7cd6b-3a10-490e-af5e-68a812d4b6f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.679274 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltqsk\" (UniqueName: \"kubernetes.io/projected/f6c7cd6b-3a10-490e-af5e-68a812d4b6f6-kube-api-access-ltqsk\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj\" (UID: \"f6c7cd6b-3a10-490e-af5e-68a812d4b6f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.780356 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6c7cd6b-3a10-490e-af5e-68a812d4b6f6-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj\" (UID: \"f6c7cd6b-3a10-490e-af5e-68a812d4b6f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.780426 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6c7cd6b-3a10-490e-af5e-68a812d4b6f6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj\" (UID: \"f6c7cd6b-3a10-490e-af5e-68a812d4b6f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.780498 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltqsk\" (UniqueName: \"kubernetes.io/projected/f6c7cd6b-3a10-490e-af5e-68a812d4b6f6-kube-api-access-ltqsk\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj\" (UID: \"f6c7cd6b-3a10-490e-af5e-68a812d4b6f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.786680 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6c7cd6b-3a10-490e-af5e-68a812d4b6f6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj\" (UID: \"f6c7cd6b-3a10-490e-af5e-68a812d4b6f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.787000 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6c7cd6b-3a10-490e-af5e-68a812d4b6f6-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj\" (UID: \"f6c7cd6b-3a10-490e-af5e-68a812d4b6f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.803817 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltqsk\" (UniqueName: \"kubernetes.io/projected/f6c7cd6b-3a10-490e-af5e-68a812d4b6f6-kube-api-access-ltqsk\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj\" (UID: \"f6c7cd6b-3a10-490e-af5e-68a812d4b6f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj" Oct 13 09:16:57 crc kubenswrapper[4685]: I1013 09:16:57.886998 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj" Oct 13 09:16:58 crc kubenswrapper[4685]: I1013 09:16:58.500960 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj"] Oct 13 09:16:59 crc kubenswrapper[4685]: I1013 09:16:59.496028 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj" event={"ID":"f6c7cd6b-3a10-490e-af5e-68a812d4b6f6","Type":"ContainerStarted","Data":"ee126614dbc9d83e910b2f8a754c43da789b616d8dab2157fd6b36e9e58630af"} Oct 13 09:17:00 crc kubenswrapper[4685]: I1013 09:17:00.510566 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj" event={"ID":"f6c7cd6b-3a10-490e-af5e-68a812d4b6f6","Type":"ContainerStarted","Data":"7f840bd2e63bef1e57196693e1d7c73bc6b5afea5e1f9e6e5c0c7f08918c193c"} Oct 13 09:17:00 crc kubenswrapper[4685]: I1013 09:17:00.533552 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj" podStartSLOduration=2.987886643 podStartE2EDuration="3.533525266s" podCreationTimestamp="2025-10-13 09:16:57 +0000 UTC" firstStartedPulling="2025-10-13 09:16:58.49363857 +0000 UTC m=+1943.641514331" lastFinishedPulling="2025-10-13 09:16:59.039277193 +0000 UTC m=+1944.187152954" observedRunningTime="2025-10-13 09:17:00.527318817 +0000 UTC m=+1945.675194678" watchObservedRunningTime="2025-10-13 09:17:00.533525266 +0000 UTC m=+1945.681401067" Oct 13 09:17:04 crc kubenswrapper[4685]: I1013 09:17:04.503134 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:17:04 crc kubenswrapper[4685]: E1013 09:17:04.503806 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:17:10 crc kubenswrapper[4685]: I1013 09:17:10.597486 4685 generic.go:334] "Generic (PLEG): container finished" podID="f6c7cd6b-3a10-490e-af5e-68a812d4b6f6" containerID="7f840bd2e63bef1e57196693e1d7c73bc6b5afea5e1f9e6e5c0c7f08918c193c" exitCode=0 Oct 13 09:17:10 crc kubenswrapper[4685]: I1013 09:17:10.597588 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj" event={"ID":"f6c7cd6b-3a10-490e-af5e-68a812d4b6f6","Type":"ContainerDied","Data":"7f840bd2e63bef1e57196693e1d7c73bc6b5afea5e1f9e6e5c0c7f08918c193c"} Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.049081 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.187712 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6c7cd6b-3a10-490e-af5e-68a812d4b6f6-ssh-key\") pod \"f6c7cd6b-3a10-490e-af5e-68a812d4b6f6\" (UID: \"f6c7cd6b-3a10-490e-af5e-68a812d4b6f6\") " Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.187839 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltqsk\" (UniqueName: \"kubernetes.io/projected/f6c7cd6b-3a10-490e-af5e-68a812d4b6f6-kube-api-access-ltqsk\") pod \"f6c7cd6b-3a10-490e-af5e-68a812d4b6f6\" (UID: \"f6c7cd6b-3a10-490e-af5e-68a812d4b6f6\") " Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.188042 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6c7cd6b-3a10-490e-af5e-68a812d4b6f6-inventory\") pod \"f6c7cd6b-3a10-490e-af5e-68a812d4b6f6\" (UID: \"f6c7cd6b-3a10-490e-af5e-68a812d4b6f6\") " Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.206129 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6c7cd6b-3a10-490e-af5e-68a812d4b6f6-kube-api-access-ltqsk" (OuterVolumeSpecName: "kube-api-access-ltqsk") pod "f6c7cd6b-3a10-490e-af5e-68a812d4b6f6" (UID: "f6c7cd6b-3a10-490e-af5e-68a812d4b6f6"). InnerVolumeSpecName "kube-api-access-ltqsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.213230 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6c7cd6b-3a10-490e-af5e-68a812d4b6f6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f6c7cd6b-3a10-490e-af5e-68a812d4b6f6" (UID: "f6c7cd6b-3a10-490e-af5e-68a812d4b6f6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.220250 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6c7cd6b-3a10-490e-af5e-68a812d4b6f6-inventory" (OuterVolumeSpecName: "inventory") pod "f6c7cd6b-3a10-490e-af5e-68a812d4b6f6" (UID: "f6c7cd6b-3a10-490e-af5e-68a812d4b6f6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.301357 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6c7cd6b-3a10-490e-af5e-68a812d4b6f6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.301406 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltqsk\" (UniqueName: \"kubernetes.io/projected/f6c7cd6b-3a10-490e-af5e-68a812d4b6f6-kube-api-access-ltqsk\") on node \"crc\" DevicePath \"\"" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.301423 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6c7cd6b-3a10-490e-af5e-68a812d4b6f6-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.618091 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj" event={"ID":"f6c7cd6b-3a10-490e-af5e-68a812d4b6f6","Type":"ContainerDied","Data":"ee126614dbc9d83e910b2f8a754c43da789b616d8dab2157fd6b36e9e58630af"} Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.618123 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.618136 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee126614dbc9d83e910b2f8a754c43da789b616d8dab2157fd6b36e9e58630af" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.715450 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f"] Oct 13 09:17:12 crc kubenswrapper[4685]: E1013 09:17:12.715978 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6c7cd6b-3a10-490e-af5e-68a812d4b6f6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.716003 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6c7cd6b-3a10-490e-af5e-68a812d4b6f6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.716241 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6c7cd6b-3a10-490e-af5e-68a812d4b6f6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.717097 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.719374 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.719577 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.721191 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.722238 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hcn4z" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.722458 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.722520 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.722484 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.722848 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.735449 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f"] Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.912704 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.913125 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.913173 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.913218 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.913256 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.913284 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.913303 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.913330 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.913388 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.913419 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.913462 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.913498 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.913515 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2798\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-kube-api-access-d2798\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:12 crc kubenswrapper[4685]: I1013 09:17:12.913560 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.015659 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.015746 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.015784 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.015825 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.015874 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.015930 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.015959 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.015980 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.016004 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.016064 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.016097 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.016143 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.016179 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.016196 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2798\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-kube-api-access-d2798\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.019372 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.021372 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.021809 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.022300 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.022428 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.023067 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.025038 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.025334 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.026338 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.026703 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.031458 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.036439 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.041322 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.043481 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2798\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-kube-api-access-d2798\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sc94f\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.331870 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.446762 4685 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.78:8081/readyz\": dial tcp 10.217.0.78:8081: connect: connection refused" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.632142 4685 generic.go:334] "Generic (PLEG): container finished" podID="b001d17a-1aea-44ba-86c5-ba6b312156c1" containerID="4e57efd2a5e99def1a73daa5f8404b15401378bfed57465d070b35fe0a552d79" exitCode=1 Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.632245 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" event={"ID":"b001d17a-1aea-44ba-86c5-ba6b312156c1","Type":"ContainerDied","Data":"4e57efd2a5e99def1a73daa5f8404b15401378bfed57465d070b35fe0a552d79"} Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.632576 4685 scope.go:117] "RemoveContainer" containerID="f4bd2015aa86f0bf7c40324d231108b8eefc9288b86f37bc304379cab4fbde84" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.633420 4685 scope.go:117] "RemoveContainer" containerID="4e57efd2a5e99def1a73daa5f8404b15401378bfed57465d070b35fe0a552d79" Oct 13 09:17:13 crc kubenswrapper[4685]: E1013 09:17:13.633709 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:17:13 crc kubenswrapper[4685]: I1013 09:17:13.931092 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f"] Oct 13 09:17:14 crc kubenswrapper[4685]: I1013 09:17:14.644903 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" event={"ID":"e8ad7931-be11-41ff-9ca5-1ad434026e44","Type":"ContainerStarted","Data":"955e7e6d340005961a65e9ae8dcfb6058756e737fa673c404473abf74d6695ee"} Oct 13 09:17:15 crc kubenswrapper[4685]: I1013 09:17:15.511164 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:17:15 crc kubenswrapper[4685]: E1013 09:17:15.511764 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:17:15 crc kubenswrapper[4685]: I1013 09:17:15.658580 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" event={"ID":"e8ad7931-be11-41ff-9ca5-1ad434026e44","Type":"ContainerStarted","Data":"a31af9c35c50d1d56edc1178bcb09f4b56e3db117db8a3fec6a9cdf9f17e3d29"} Oct 13 09:17:15 crc kubenswrapper[4685]: I1013 09:17:15.688496 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" podStartSLOduration=3.232935074 podStartE2EDuration="3.688477313s" podCreationTimestamp="2025-10-13 09:17:12 +0000 UTC" firstStartedPulling="2025-10-13 09:17:13.954109289 +0000 UTC m=+1959.101985050" lastFinishedPulling="2025-10-13 09:17:14.409651528 +0000 UTC m=+1959.557527289" observedRunningTime="2025-10-13 09:17:15.684239148 +0000 UTC m=+1960.832114919" watchObservedRunningTime="2025-10-13 09:17:15.688477313 +0000 UTC m=+1960.836353084" Oct 13 09:17:23 crc kubenswrapper[4685]: I1013 09:17:23.442781 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:17:23 crc kubenswrapper[4685]: I1013 09:17:23.443307 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:17:23 crc kubenswrapper[4685]: I1013 09:17:23.443989 4685 scope.go:117] "RemoveContainer" containerID="4e57efd2a5e99def1a73daa5f8404b15401378bfed57465d070b35fe0a552d79" Oct 13 09:17:23 crc kubenswrapper[4685]: E1013 09:17:23.444232 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:17:30 crc kubenswrapper[4685]: I1013 09:17:30.502839 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:17:30 crc kubenswrapper[4685]: E1013 09:17:30.503731 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:17:38 crc kubenswrapper[4685]: I1013 09:17:38.502629 4685 scope.go:117] "RemoveContainer" containerID="4e57efd2a5e99def1a73daa5f8404b15401378bfed57465d070b35fe0a552d79" Oct 13 09:17:38 crc kubenswrapper[4685]: E1013 09:17:38.504348 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:17:45 crc kubenswrapper[4685]: I1013 09:17:45.507683 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:17:45 crc kubenswrapper[4685]: E1013 09:17:45.508479 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:17:50 crc kubenswrapper[4685]: I1013 09:17:50.502270 4685 scope.go:117] "RemoveContainer" containerID="4e57efd2a5e99def1a73daa5f8404b15401378bfed57465d070b35fe0a552d79" Oct 13 09:17:50 crc kubenswrapper[4685]: E1013 09:17:50.502762 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:17:54 crc kubenswrapper[4685]: I1013 09:17:54.465878 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m6j5f"] Oct 13 09:17:54 crc kubenswrapper[4685]: I1013 09:17:54.470160 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m6j5f" Oct 13 09:17:54 crc kubenswrapper[4685]: I1013 09:17:54.481831 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m6j5f"] Oct 13 09:17:54 crc kubenswrapper[4685]: I1013 09:17:54.486148 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed00208c-afc6-4bc6-89b4-d23b93969110-utilities\") pod \"community-operators-m6j5f\" (UID: \"ed00208c-afc6-4bc6-89b4-d23b93969110\") " pod="openshift-marketplace/community-operators-m6j5f" Oct 13 09:17:54 crc kubenswrapper[4685]: I1013 09:17:54.486274 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt7jz\" (UniqueName: \"kubernetes.io/projected/ed00208c-afc6-4bc6-89b4-d23b93969110-kube-api-access-mt7jz\") pod \"community-operators-m6j5f\" (UID: \"ed00208c-afc6-4bc6-89b4-d23b93969110\") " pod="openshift-marketplace/community-operators-m6j5f" Oct 13 09:17:54 crc kubenswrapper[4685]: I1013 09:17:54.486422 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed00208c-afc6-4bc6-89b4-d23b93969110-catalog-content\") pod \"community-operators-m6j5f\" (UID: \"ed00208c-afc6-4bc6-89b4-d23b93969110\") " pod="openshift-marketplace/community-operators-m6j5f" Oct 13 09:17:54 crc kubenswrapper[4685]: I1013 09:17:54.588891 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed00208c-afc6-4bc6-89b4-d23b93969110-utilities\") pod \"community-operators-m6j5f\" (UID: \"ed00208c-afc6-4bc6-89b4-d23b93969110\") " pod="openshift-marketplace/community-operators-m6j5f" Oct 13 09:17:54 crc kubenswrapper[4685]: I1013 09:17:54.589041 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt7jz\" (UniqueName: \"kubernetes.io/projected/ed00208c-afc6-4bc6-89b4-d23b93969110-kube-api-access-mt7jz\") pod \"community-operators-m6j5f\" (UID: \"ed00208c-afc6-4bc6-89b4-d23b93969110\") " pod="openshift-marketplace/community-operators-m6j5f" Oct 13 09:17:54 crc kubenswrapper[4685]: I1013 09:17:54.589163 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed00208c-afc6-4bc6-89b4-d23b93969110-catalog-content\") pod \"community-operators-m6j5f\" (UID: \"ed00208c-afc6-4bc6-89b4-d23b93969110\") " pod="openshift-marketplace/community-operators-m6j5f" Oct 13 09:17:54 crc kubenswrapper[4685]: I1013 09:17:54.589960 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed00208c-afc6-4bc6-89b4-d23b93969110-utilities\") pod \"community-operators-m6j5f\" (UID: \"ed00208c-afc6-4bc6-89b4-d23b93969110\") " pod="openshift-marketplace/community-operators-m6j5f" Oct 13 09:17:54 crc kubenswrapper[4685]: I1013 09:17:54.590103 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed00208c-afc6-4bc6-89b4-d23b93969110-catalog-content\") pod \"community-operators-m6j5f\" (UID: \"ed00208c-afc6-4bc6-89b4-d23b93969110\") " pod="openshift-marketplace/community-operators-m6j5f" Oct 13 09:17:54 crc kubenswrapper[4685]: I1013 09:17:54.619193 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt7jz\" (UniqueName: \"kubernetes.io/projected/ed00208c-afc6-4bc6-89b4-d23b93969110-kube-api-access-mt7jz\") pod \"community-operators-m6j5f\" (UID: \"ed00208c-afc6-4bc6-89b4-d23b93969110\") " pod="openshift-marketplace/community-operators-m6j5f" Oct 13 09:17:54 crc kubenswrapper[4685]: I1013 09:17:54.794863 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m6j5f" Oct 13 09:17:55 crc kubenswrapper[4685]: I1013 09:17:55.388882 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m6j5f"] Oct 13 09:17:55 crc kubenswrapper[4685]: W1013 09:17:55.390509 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded00208c_afc6_4bc6_89b4_d23b93969110.slice/crio-460066f75563c69b66cfdcd77f220c52092f1d9c327bf4988d91fe5acabc9529 WatchSource:0}: Error finding container 460066f75563c69b66cfdcd77f220c52092f1d9c327bf4988d91fe5acabc9529: Status 404 returned error can't find the container with id 460066f75563c69b66cfdcd77f220c52092f1d9c327bf4988d91fe5acabc9529 Oct 13 09:17:56 crc kubenswrapper[4685]: I1013 09:17:56.043580 4685 generic.go:334] "Generic (PLEG): container finished" podID="ed00208c-afc6-4bc6-89b4-d23b93969110" containerID="bc6eee3ccbd4d2a32b3256d9a5a9c6a2992ae2679d9365bb078b40dbed3157bd" exitCode=0 Oct 13 09:17:56 crc kubenswrapper[4685]: I1013 09:17:56.043636 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m6j5f" event={"ID":"ed00208c-afc6-4bc6-89b4-d23b93969110","Type":"ContainerDied","Data":"bc6eee3ccbd4d2a32b3256d9a5a9c6a2992ae2679d9365bb078b40dbed3157bd"} Oct 13 09:17:56 crc kubenswrapper[4685]: I1013 09:17:56.043892 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m6j5f" event={"ID":"ed00208c-afc6-4bc6-89b4-d23b93969110","Type":"ContainerStarted","Data":"460066f75563c69b66cfdcd77f220c52092f1d9c327bf4988d91fe5acabc9529"} Oct 13 09:17:56 crc kubenswrapper[4685]: I1013 09:17:56.047602 4685 generic.go:334] "Generic (PLEG): container finished" podID="e8ad7931-be11-41ff-9ca5-1ad434026e44" containerID="a31af9c35c50d1d56edc1178bcb09f4b56e3db117db8a3fec6a9cdf9f17e3d29" exitCode=0 Oct 13 09:17:56 crc kubenswrapper[4685]: I1013 09:17:56.047707 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" event={"ID":"e8ad7931-be11-41ff-9ca5-1ad434026e44","Type":"ContainerDied","Data":"a31af9c35c50d1d56edc1178bcb09f4b56e3db117db8a3fec6a9cdf9f17e3d29"} Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.058300 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m6j5f" event={"ID":"ed00208c-afc6-4bc6-89b4-d23b93969110","Type":"ContainerStarted","Data":"e3306f1a8f0dd9471d91f81a1476235ca8da727f4982d14fc22d2118ab606fad"} Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.456391 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.644017 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"e8ad7931-be11-41ff-9ca5-1ad434026e44\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.644864 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"e8ad7931-be11-41ff-9ca5-1ad434026e44\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.645766 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-repo-setup-combined-ca-bundle\") pod \"e8ad7931-be11-41ff-9ca5-1ad434026e44\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.646007 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-libvirt-combined-ca-bundle\") pod \"e8ad7931-be11-41ff-9ca5-1ad434026e44\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.646205 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"e8ad7931-be11-41ff-9ca5-1ad434026e44\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.646411 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-nova-combined-ca-bundle\") pod \"e8ad7931-be11-41ff-9ca5-1ad434026e44\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.646560 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-bootstrap-combined-ca-bundle\") pod \"e8ad7931-be11-41ff-9ca5-1ad434026e44\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.646732 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2798\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-kube-api-access-d2798\") pod \"e8ad7931-be11-41ff-9ca5-1ad434026e44\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.646890 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-telemetry-combined-ca-bundle\") pod \"e8ad7931-be11-41ff-9ca5-1ad434026e44\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.647245 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-ovn-combined-ca-bundle\") pod \"e8ad7931-be11-41ff-9ca5-1ad434026e44\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.647442 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-ovn-default-certs-0\") pod \"e8ad7931-be11-41ff-9ca5-1ad434026e44\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.647634 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-inventory\") pod \"e8ad7931-be11-41ff-9ca5-1ad434026e44\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.648441 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-neutron-metadata-combined-ca-bundle\") pod \"e8ad7931-be11-41ff-9ca5-1ad434026e44\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.648894 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-ssh-key\") pod \"e8ad7931-be11-41ff-9ca5-1ad434026e44\" (UID: \"e8ad7931-be11-41ff-9ca5-1ad434026e44\") " Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.650013 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "e8ad7931-be11-41ff-9ca5-1ad434026e44" (UID: "e8ad7931-be11-41ff-9ca5-1ad434026e44"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.651013 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "e8ad7931-be11-41ff-9ca5-1ad434026e44" (UID: "e8ad7931-be11-41ff-9ca5-1ad434026e44"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.651119 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "e8ad7931-be11-41ff-9ca5-1ad434026e44" (UID: "e8ad7931-be11-41ff-9ca5-1ad434026e44"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.651396 4685 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.651557 4685 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.651697 4685 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.653519 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "e8ad7931-be11-41ff-9ca5-1ad434026e44" (UID: "e8ad7931-be11-41ff-9ca5-1ad434026e44"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.654397 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "e8ad7931-be11-41ff-9ca5-1ad434026e44" (UID: "e8ad7931-be11-41ff-9ca5-1ad434026e44"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.655015 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "e8ad7931-be11-41ff-9ca5-1ad434026e44" (UID: "e8ad7931-be11-41ff-9ca5-1ad434026e44"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.657052 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "e8ad7931-be11-41ff-9ca5-1ad434026e44" (UID: "e8ad7931-be11-41ff-9ca5-1ad434026e44"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.657330 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "e8ad7931-be11-41ff-9ca5-1ad434026e44" (UID: "e8ad7931-be11-41ff-9ca5-1ad434026e44"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.657413 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "e8ad7931-be11-41ff-9ca5-1ad434026e44" (UID: "e8ad7931-be11-41ff-9ca5-1ad434026e44"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.657510 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-kube-api-access-d2798" (OuterVolumeSpecName: "kube-api-access-d2798") pod "e8ad7931-be11-41ff-9ca5-1ad434026e44" (UID: "e8ad7931-be11-41ff-9ca5-1ad434026e44"). InnerVolumeSpecName "kube-api-access-d2798". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.657770 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "e8ad7931-be11-41ff-9ca5-1ad434026e44" (UID: "e8ad7931-be11-41ff-9ca5-1ad434026e44"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.660359 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "e8ad7931-be11-41ff-9ca5-1ad434026e44" (UID: "e8ad7931-be11-41ff-9ca5-1ad434026e44"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.678379 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-inventory" (OuterVolumeSpecName: "inventory") pod "e8ad7931-be11-41ff-9ca5-1ad434026e44" (UID: "e8ad7931-be11-41ff-9ca5-1ad434026e44"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.679890 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e8ad7931-be11-41ff-9ca5-1ad434026e44" (UID: "e8ad7931-be11-41ff-9ca5-1ad434026e44"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.753802 4685 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.754068 4685 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.754185 4685 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.754315 4685 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.754403 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2798\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-kube-api-access-d2798\") on node \"crc\" DevicePath \"\"" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.754477 4685 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.754555 4685 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.754630 4685 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e8ad7931-be11-41ff-9ca5-1ad434026e44-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.754705 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.754789 4685 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:17:57 crc kubenswrapper[4685]: I1013 09:17:57.754867 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e8ad7931-be11-41ff-9ca5-1ad434026e44-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.068609 4685 generic.go:334] "Generic (PLEG): container finished" podID="ed00208c-afc6-4bc6-89b4-d23b93969110" containerID="e3306f1a8f0dd9471d91f81a1476235ca8da727f4982d14fc22d2118ab606fad" exitCode=0 Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.068696 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m6j5f" event={"ID":"ed00208c-afc6-4bc6-89b4-d23b93969110","Type":"ContainerDied","Data":"e3306f1a8f0dd9471d91f81a1476235ca8da727f4982d14fc22d2118ab606fad"} Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.073161 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" event={"ID":"e8ad7931-be11-41ff-9ca5-1ad434026e44","Type":"ContainerDied","Data":"955e7e6d340005961a65e9ae8dcfb6058756e737fa673c404473abf74d6695ee"} Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.073191 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="955e7e6d340005961a65e9ae8dcfb6058756e737fa673c404473abf74d6695ee" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.073240 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sc94f" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.195322 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn"] Oct 13 09:17:58 crc kubenswrapper[4685]: E1013 09:17:58.195791 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ad7931-be11-41ff-9ca5-1ad434026e44" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.195809 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ad7931-be11-41ff-9ca5-1ad434026e44" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.196098 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8ad7931-be11-41ff-9ca5-1ad434026e44" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.196744 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.200782 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hcn4z" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.200800 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.201267 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.201483 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.205595 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn"] Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.211098 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.265104 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57483ac7-dfa4-4538-8605-11b9215432dc-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nwnvn\" (UID: \"57483ac7-dfa4-4538-8605-11b9215432dc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.265178 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/57483ac7-dfa4-4538-8605-11b9215432dc-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nwnvn\" (UID: \"57483ac7-dfa4-4538-8605-11b9215432dc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.265239 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pbdx\" (UniqueName: \"kubernetes.io/projected/57483ac7-dfa4-4538-8605-11b9215432dc-kube-api-access-8pbdx\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nwnvn\" (UID: \"57483ac7-dfa4-4538-8605-11b9215432dc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.265349 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57483ac7-dfa4-4538-8605-11b9215432dc-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nwnvn\" (UID: \"57483ac7-dfa4-4538-8605-11b9215432dc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.265496 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57483ac7-dfa4-4538-8605-11b9215432dc-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nwnvn\" (UID: \"57483ac7-dfa4-4538-8605-11b9215432dc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.367223 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/57483ac7-dfa4-4538-8605-11b9215432dc-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nwnvn\" (UID: \"57483ac7-dfa4-4538-8605-11b9215432dc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.367307 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pbdx\" (UniqueName: \"kubernetes.io/projected/57483ac7-dfa4-4538-8605-11b9215432dc-kube-api-access-8pbdx\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nwnvn\" (UID: \"57483ac7-dfa4-4538-8605-11b9215432dc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.367340 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57483ac7-dfa4-4538-8605-11b9215432dc-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nwnvn\" (UID: \"57483ac7-dfa4-4538-8605-11b9215432dc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.367402 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57483ac7-dfa4-4538-8605-11b9215432dc-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nwnvn\" (UID: \"57483ac7-dfa4-4538-8605-11b9215432dc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.367500 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57483ac7-dfa4-4538-8605-11b9215432dc-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nwnvn\" (UID: \"57483ac7-dfa4-4538-8605-11b9215432dc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.368751 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/57483ac7-dfa4-4538-8605-11b9215432dc-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nwnvn\" (UID: \"57483ac7-dfa4-4538-8605-11b9215432dc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.374809 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57483ac7-dfa4-4538-8605-11b9215432dc-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nwnvn\" (UID: \"57483ac7-dfa4-4538-8605-11b9215432dc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.375171 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57483ac7-dfa4-4538-8605-11b9215432dc-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nwnvn\" (UID: \"57483ac7-dfa4-4538-8605-11b9215432dc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.376608 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57483ac7-dfa4-4538-8605-11b9215432dc-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nwnvn\" (UID: \"57483ac7-dfa4-4538-8605-11b9215432dc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.388051 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pbdx\" (UniqueName: \"kubernetes.io/projected/57483ac7-dfa4-4538-8605-11b9215432dc-kube-api-access-8pbdx\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nwnvn\" (UID: \"57483ac7-dfa4-4538-8605-11b9215432dc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" Oct 13 09:17:58 crc kubenswrapper[4685]: I1013 09:17:58.514842 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" Oct 13 09:17:59 crc kubenswrapper[4685]: I1013 09:17:59.083998 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn"] Oct 13 09:17:59 crc kubenswrapper[4685]: I1013 09:17:59.086123 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m6j5f" event={"ID":"ed00208c-afc6-4bc6-89b4-d23b93969110","Type":"ContainerStarted","Data":"b3b5ae602fee5471a7fe50bc65af42b670c5bc8ae23a45ed34dfc86617ba003d"} Oct 13 09:17:59 crc kubenswrapper[4685]: W1013 09:17:59.092674 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57483ac7_dfa4_4538_8605_11b9215432dc.slice/crio-360b3550993e3691b78805643b521474ab2683ad9e5ce60af1af80af91ba1e70 WatchSource:0}: Error finding container 360b3550993e3691b78805643b521474ab2683ad9e5ce60af1af80af91ba1e70: Status 404 returned error can't find the container with id 360b3550993e3691b78805643b521474ab2683ad9e5ce60af1af80af91ba1e70 Oct 13 09:17:59 crc kubenswrapper[4685]: I1013 09:17:59.111055 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m6j5f" podStartSLOduration=2.651820544 podStartE2EDuration="5.111028682s" podCreationTimestamp="2025-10-13 09:17:54 +0000 UTC" firstStartedPulling="2025-10-13 09:17:56.046034033 +0000 UTC m=+2001.193909794" lastFinishedPulling="2025-10-13 09:17:58.505242171 +0000 UTC m=+2003.653117932" observedRunningTime="2025-10-13 09:17:59.104129895 +0000 UTC m=+2004.252005666" watchObservedRunningTime="2025-10-13 09:17:59.111028682 +0000 UTC m=+2004.258904443" Oct 13 09:17:59 crc kubenswrapper[4685]: I1013 09:17:59.502847 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:18:00 crc kubenswrapper[4685]: I1013 09:18:00.094585 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" event={"ID":"57483ac7-dfa4-4538-8605-11b9215432dc","Type":"ContainerStarted","Data":"f22893f68a2d2c3e5c1282b19433866bd1188a93244ae5750ddd82a80e09f5a1"} Oct 13 09:18:00 crc kubenswrapper[4685]: I1013 09:18:00.095222 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" event={"ID":"57483ac7-dfa4-4538-8605-11b9215432dc","Type":"ContainerStarted","Data":"360b3550993e3691b78805643b521474ab2683ad9e5ce60af1af80af91ba1e70"} Oct 13 09:18:00 crc kubenswrapper[4685]: I1013 09:18:00.099164 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerStarted","Data":"1c5982a8f152978248be74b1441fee44fe6cefb88272b0a7d3055c30ef13a327"} Oct 13 09:18:00 crc kubenswrapper[4685]: I1013 09:18:00.113875 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" podStartSLOduration=1.623485611 podStartE2EDuration="2.113860109s" podCreationTimestamp="2025-10-13 09:17:58 +0000 UTC" firstStartedPulling="2025-10-13 09:17:59.096144297 +0000 UTC m=+2004.244020058" lastFinishedPulling="2025-10-13 09:17:59.586518795 +0000 UTC m=+2004.734394556" observedRunningTime="2025-10-13 09:18:00.11167868 +0000 UTC m=+2005.259554451" watchObservedRunningTime="2025-10-13 09:18:00.113860109 +0000 UTC m=+2005.261735870" Oct 13 09:18:01 crc kubenswrapper[4685]: I1013 09:18:01.505196 4685 scope.go:117] "RemoveContainer" containerID="4e57efd2a5e99def1a73daa5f8404b15401378bfed57465d070b35fe0a552d79" Oct 13 09:18:01 crc kubenswrapper[4685]: E1013 09:18:01.506126 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:18:03 crc kubenswrapper[4685]: I1013 09:18:03.274847 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fv4xw"] Oct 13 09:18:03 crc kubenswrapper[4685]: I1013 09:18:03.277129 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fv4xw" Oct 13 09:18:03 crc kubenswrapper[4685]: I1013 09:18:03.290540 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fv4xw"] Oct 13 09:18:03 crc kubenswrapper[4685]: I1013 09:18:03.368987 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/810071be-34b5-4629-b47a-17a3bb0533dd-utilities\") pod \"certified-operators-fv4xw\" (UID: \"810071be-34b5-4629-b47a-17a3bb0533dd\") " pod="openshift-marketplace/certified-operators-fv4xw" Oct 13 09:18:03 crc kubenswrapper[4685]: I1013 09:18:03.369100 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/810071be-34b5-4629-b47a-17a3bb0533dd-catalog-content\") pod \"certified-operators-fv4xw\" (UID: \"810071be-34b5-4629-b47a-17a3bb0533dd\") " pod="openshift-marketplace/certified-operators-fv4xw" Oct 13 09:18:03 crc kubenswrapper[4685]: I1013 09:18:03.369174 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqfv7\" (UniqueName: \"kubernetes.io/projected/810071be-34b5-4629-b47a-17a3bb0533dd-kube-api-access-jqfv7\") pod \"certified-operators-fv4xw\" (UID: \"810071be-34b5-4629-b47a-17a3bb0533dd\") " pod="openshift-marketplace/certified-operators-fv4xw" Oct 13 09:18:03 crc kubenswrapper[4685]: I1013 09:18:03.470667 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/810071be-34b5-4629-b47a-17a3bb0533dd-utilities\") pod \"certified-operators-fv4xw\" (UID: \"810071be-34b5-4629-b47a-17a3bb0533dd\") " pod="openshift-marketplace/certified-operators-fv4xw" Oct 13 09:18:03 crc kubenswrapper[4685]: I1013 09:18:03.470781 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/810071be-34b5-4629-b47a-17a3bb0533dd-catalog-content\") pod \"certified-operators-fv4xw\" (UID: \"810071be-34b5-4629-b47a-17a3bb0533dd\") " pod="openshift-marketplace/certified-operators-fv4xw" Oct 13 09:18:03 crc kubenswrapper[4685]: I1013 09:18:03.470847 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqfv7\" (UniqueName: \"kubernetes.io/projected/810071be-34b5-4629-b47a-17a3bb0533dd-kube-api-access-jqfv7\") pod \"certified-operators-fv4xw\" (UID: \"810071be-34b5-4629-b47a-17a3bb0533dd\") " pod="openshift-marketplace/certified-operators-fv4xw" Oct 13 09:18:03 crc kubenswrapper[4685]: I1013 09:18:03.471426 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/810071be-34b5-4629-b47a-17a3bb0533dd-utilities\") pod \"certified-operators-fv4xw\" (UID: \"810071be-34b5-4629-b47a-17a3bb0533dd\") " pod="openshift-marketplace/certified-operators-fv4xw" Oct 13 09:18:03 crc kubenswrapper[4685]: I1013 09:18:03.471753 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/810071be-34b5-4629-b47a-17a3bb0533dd-catalog-content\") pod \"certified-operators-fv4xw\" (UID: \"810071be-34b5-4629-b47a-17a3bb0533dd\") " pod="openshift-marketplace/certified-operators-fv4xw" Oct 13 09:18:03 crc kubenswrapper[4685]: I1013 09:18:03.508108 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqfv7\" (UniqueName: \"kubernetes.io/projected/810071be-34b5-4629-b47a-17a3bb0533dd-kube-api-access-jqfv7\") pod \"certified-operators-fv4xw\" (UID: \"810071be-34b5-4629-b47a-17a3bb0533dd\") " pod="openshift-marketplace/certified-operators-fv4xw" Oct 13 09:18:03 crc kubenswrapper[4685]: I1013 09:18:03.594826 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fv4xw" Oct 13 09:18:04 crc kubenswrapper[4685]: I1013 09:18:04.120483 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fv4xw"] Oct 13 09:18:04 crc kubenswrapper[4685]: I1013 09:18:04.795068 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m6j5f" Oct 13 09:18:04 crc kubenswrapper[4685]: I1013 09:18:04.795629 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m6j5f" Oct 13 09:18:04 crc kubenswrapper[4685]: I1013 09:18:04.857159 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m6j5f" Oct 13 09:18:05 crc kubenswrapper[4685]: I1013 09:18:05.137382 4685 generic.go:334] "Generic (PLEG): container finished" podID="810071be-34b5-4629-b47a-17a3bb0533dd" containerID="75d572c74f7d65e4650127afc71c4b9f87768c8a1b8f8655af95010832aac900" exitCode=0 Oct 13 09:18:05 crc kubenswrapper[4685]: I1013 09:18:05.138115 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv4xw" event={"ID":"810071be-34b5-4629-b47a-17a3bb0533dd","Type":"ContainerDied","Data":"75d572c74f7d65e4650127afc71c4b9f87768c8a1b8f8655af95010832aac900"} Oct 13 09:18:05 crc kubenswrapper[4685]: I1013 09:18:05.138159 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv4xw" event={"ID":"810071be-34b5-4629-b47a-17a3bb0533dd","Type":"ContainerStarted","Data":"81d37ace3d759d0c7ca41fd9dbfeff3eb4a39f786f6d38f67a9b696906c67e1b"} Oct 13 09:18:05 crc kubenswrapper[4685]: I1013 09:18:05.203269 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m6j5f" Oct 13 09:18:07 crc kubenswrapper[4685]: I1013 09:18:07.159788 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv4xw" event={"ID":"810071be-34b5-4629-b47a-17a3bb0533dd","Type":"ContainerStarted","Data":"a08c3032469671c4432e0cbeecde7ed2eb22523e011e5bc421c039c34a26c622"} Oct 13 09:18:07 crc kubenswrapper[4685]: I1013 09:18:07.246870 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m6j5f"] Oct 13 09:18:07 crc kubenswrapper[4685]: I1013 09:18:07.247097 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-m6j5f" podUID="ed00208c-afc6-4bc6-89b4-d23b93969110" containerName="registry-server" containerID="cri-o://b3b5ae602fee5471a7fe50bc65af42b670c5bc8ae23a45ed34dfc86617ba003d" gracePeriod=2 Oct 13 09:18:07 crc kubenswrapper[4685]: I1013 09:18:07.716291 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m6j5f" Oct 13 09:18:07 crc kubenswrapper[4685]: I1013 09:18:07.878864 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mt7jz\" (UniqueName: \"kubernetes.io/projected/ed00208c-afc6-4bc6-89b4-d23b93969110-kube-api-access-mt7jz\") pod \"ed00208c-afc6-4bc6-89b4-d23b93969110\" (UID: \"ed00208c-afc6-4bc6-89b4-d23b93969110\") " Oct 13 09:18:07 crc kubenswrapper[4685]: I1013 09:18:07.879070 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed00208c-afc6-4bc6-89b4-d23b93969110-utilities\") pod \"ed00208c-afc6-4bc6-89b4-d23b93969110\" (UID: \"ed00208c-afc6-4bc6-89b4-d23b93969110\") " Oct 13 09:18:07 crc kubenswrapper[4685]: I1013 09:18:07.879107 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed00208c-afc6-4bc6-89b4-d23b93969110-catalog-content\") pod \"ed00208c-afc6-4bc6-89b4-d23b93969110\" (UID: \"ed00208c-afc6-4bc6-89b4-d23b93969110\") " Oct 13 09:18:07 crc kubenswrapper[4685]: I1013 09:18:07.880211 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed00208c-afc6-4bc6-89b4-d23b93969110-utilities" (OuterVolumeSpecName: "utilities") pod "ed00208c-afc6-4bc6-89b4-d23b93969110" (UID: "ed00208c-afc6-4bc6-89b4-d23b93969110"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:18:07 crc kubenswrapper[4685]: I1013 09:18:07.888750 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed00208c-afc6-4bc6-89b4-d23b93969110-kube-api-access-mt7jz" (OuterVolumeSpecName: "kube-api-access-mt7jz") pod "ed00208c-afc6-4bc6-89b4-d23b93969110" (UID: "ed00208c-afc6-4bc6-89b4-d23b93969110"). InnerVolumeSpecName "kube-api-access-mt7jz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:18:07 crc kubenswrapper[4685]: I1013 09:18:07.926159 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed00208c-afc6-4bc6-89b4-d23b93969110-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ed00208c-afc6-4bc6-89b4-d23b93969110" (UID: "ed00208c-afc6-4bc6-89b4-d23b93969110"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:18:07 crc kubenswrapper[4685]: I1013 09:18:07.981805 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed00208c-afc6-4bc6-89b4-d23b93969110-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 09:18:07 crc kubenswrapper[4685]: I1013 09:18:07.981853 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed00208c-afc6-4bc6-89b4-d23b93969110-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 09:18:07 crc kubenswrapper[4685]: I1013 09:18:07.981866 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mt7jz\" (UniqueName: \"kubernetes.io/projected/ed00208c-afc6-4bc6-89b4-d23b93969110-kube-api-access-mt7jz\") on node \"crc\" DevicePath \"\"" Oct 13 09:18:08 crc kubenswrapper[4685]: I1013 09:18:08.173098 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m6j5f" Oct 13 09:18:08 crc kubenswrapper[4685]: I1013 09:18:08.173121 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m6j5f" event={"ID":"ed00208c-afc6-4bc6-89b4-d23b93969110","Type":"ContainerDied","Data":"b3b5ae602fee5471a7fe50bc65af42b670c5bc8ae23a45ed34dfc86617ba003d"} Oct 13 09:18:08 crc kubenswrapper[4685]: I1013 09:18:08.173162 4685 scope.go:117] "RemoveContainer" containerID="b3b5ae602fee5471a7fe50bc65af42b670c5bc8ae23a45ed34dfc86617ba003d" Oct 13 09:18:08 crc kubenswrapper[4685]: I1013 09:18:08.173052 4685 generic.go:334] "Generic (PLEG): container finished" podID="ed00208c-afc6-4bc6-89b4-d23b93969110" containerID="b3b5ae602fee5471a7fe50bc65af42b670c5bc8ae23a45ed34dfc86617ba003d" exitCode=0 Oct 13 09:18:08 crc kubenswrapper[4685]: I1013 09:18:08.173359 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m6j5f" event={"ID":"ed00208c-afc6-4bc6-89b4-d23b93969110","Type":"ContainerDied","Data":"460066f75563c69b66cfdcd77f220c52092f1d9c327bf4988d91fe5acabc9529"} Oct 13 09:18:08 crc kubenswrapper[4685]: I1013 09:18:08.215685 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m6j5f"] Oct 13 09:18:08 crc kubenswrapper[4685]: I1013 09:18:08.216818 4685 scope.go:117] "RemoveContainer" containerID="e3306f1a8f0dd9471d91f81a1476235ca8da727f4982d14fc22d2118ab606fad" Oct 13 09:18:08 crc kubenswrapper[4685]: I1013 09:18:08.230366 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-m6j5f"] Oct 13 09:18:08 crc kubenswrapper[4685]: I1013 09:18:08.235186 4685 scope.go:117] "RemoveContainer" containerID="bc6eee3ccbd4d2a32b3256d9a5a9c6a2992ae2679d9365bb078b40dbed3157bd" Oct 13 09:18:08 crc kubenswrapper[4685]: I1013 09:18:08.337175 4685 scope.go:117] "RemoveContainer" containerID="b3b5ae602fee5471a7fe50bc65af42b670c5bc8ae23a45ed34dfc86617ba003d" Oct 13 09:18:08 crc kubenswrapper[4685]: E1013 09:18:08.337647 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3b5ae602fee5471a7fe50bc65af42b670c5bc8ae23a45ed34dfc86617ba003d\": container with ID starting with b3b5ae602fee5471a7fe50bc65af42b670c5bc8ae23a45ed34dfc86617ba003d not found: ID does not exist" containerID="b3b5ae602fee5471a7fe50bc65af42b670c5bc8ae23a45ed34dfc86617ba003d" Oct 13 09:18:08 crc kubenswrapper[4685]: I1013 09:18:08.337698 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3b5ae602fee5471a7fe50bc65af42b670c5bc8ae23a45ed34dfc86617ba003d"} err="failed to get container status \"b3b5ae602fee5471a7fe50bc65af42b670c5bc8ae23a45ed34dfc86617ba003d\": rpc error: code = NotFound desc = could not find container \"b3b5ae602fee5471a7fe50bc65af42b670c5bc8ae23a45ed34dfc86617ba003d\": container with ID starting with b3b5ae602fee5471a7fe50bc65af42b670c5bc8ae23a45ed34dfc86617ba003d not found: ID does not exist" Oct 13 09:18:08 crc kubenswrapper[4685]: I1013 09:18:08.337730 4685 scope.go:117] "RemoveContainer" containerID="e3306f1a8f0dd9471d91f81a1476235ca8da727f4982d14fc22d2118ab606fad" Oct 13 09:18:08 crc kubenswrapper[4685]: E1013 09:18:08.338079 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3306f1a8f0dd9471d91f81a1476235ca8da727f4982d14fc22d2118ab606fad\": container with ID starting with e3306f1a8f0dd9471d91f81a1476235ca8da727f4982d14fc22d2118ab606fad not found: ID does not exist" containerID="e3306f1a8f0dd9471d91f81a1476235ca8da727f4982d14fc22d2118ab606fad" Oct 13 09:18:08 crc kubenswrapper[4685]: I1013 09:18:08.338109 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3306f1a8f0dd9471d91f81a1476235ca8da727f4982d14fc22d2118ab606fad"} err="failed to get container status \"e3306f1a8f0dd9471d91f81a1476235ca8da727f4982d14fc22d2118ab606fad\": rpc error: code = NotFound desc = could not find container \"e3306f1a8f0dd9471d91f81a1476235ca8da727f4982d14fc22d2118ab606fad\": container with ID starting with e3306f1a8f0dd9471d91f81a1476235ca8da727f4982d14fc22d2118ab606fad not found: ID does not exist" Oct 13 09:18:08 crc kubenswrapper[4685]: I1013 09:18:08.338130 4685 scope.go:117] "RemoveContainer" containerID="bc6eee3ccbd4d2a32b3256d9a5a9c6a2992ae2679d9365bb078b40dbed3157bd" Oct 13 09:18:08 crc kubenswrapper[4685]: E1013 09:18:08.338417 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc6eee3ccbd4d2a32b3256d9a5a9c6a2992ae2679d9365bb078b40dbed3157bd\": container with ID starting with bc6eee3ccbd4d2a32b3256d9a5a9c6a2992ae2679d9365bb078b40dbed3157bd not found: ID does not exist" containerID="bc6eee3ccbd4d2a32b3256d9a5a9c6a2992ae2679d9365bb078b40dbed3157bd" Oct 13 09:18:08 crc kubenswrapper[4685]: I1013 09:18:08.338444 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc6eee3ccbd4d2a32b3256d9a5a9c6a2992ae2679d9365bb078b40dbed3157bd"} err="failed to get container status \"bc6eee3ccbd4d2a32b3256d9a5a9c6a2992ae2679d9365bb078b40dbed3157bd\": rpc error: code = NotFound desc = could not find container \"bc6eee3ccbd4d2a32b3256d9a5a9c6a2992ae2679d9365bb078b40dbed3157bd\": container with ID starting with bc6eee3ccbd4d2a32b3256d9a5a9c6a2992ae2679d9365bb078b40dbed3157bd not found: ID does not exist" Oct 13 09:18:09 crc kubenswrapper[4685]: I1013 09:18:09.189316 4685 generic.go:334] "Generic (PLEG): container finished" podID="810071be-34b5-4629-b47a-17a3bb0533dd" containerID="a08c3032469671c4432e0cbeecde7ed2eb22523e011e5bc421c039c34a26c622" exitCode=0 Oct 13 09:18:09 crc kubenswrapper[4685]: I1013 09:18:09.189374 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv4xw" event={"ID":"810071be-34b5-4629-b47a-17a3bb0533dd","Type":"ContainerDied","Data":"a08c3032469671c4432e0cbeecde7ed2eb22523e011e5bc421c039c34a26c622"} Oct 13 09:18:09 crc kubenswrapper[4685]: I1013 09:18:09.535977 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed00208c-afc6-4bc6-89b4-d23b93969110" path="/var/lib/kubelet/pods/ed00208c-afc6-4bc6-89b4-d23b93969110/volumes" Oct 13 09:18:10 crc kubenswrapper[4685]: I1013 09:18:10.200230 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv4xw" event={"ID":"810071be-34b5-4629-b47a-17a3bb0533dd","Type":"ContainerStarted","Data":"a45b788714e9ebd35f78c05f6a2b6447098e9ecac6b956e62dbafdbe0cd32071"} Oct 13 09:18:10 crc kubenswrapper[4685]: I1013 09:18:10.220518 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fv4xw" podStartSLOduration=2.65175581 podStartE2EDuration="7.220493152s" podCreationTimestamp="2025-10-13 09:18:03 +0000 UTC" firstStartedPulling="2025-10-13 09:18:05.140445893 +0000 UTC m=+2010.288321654" lastFinishedPulling="2025-10-13 09:18:09.709183235 +0000 UTC m=+2014.857058996" observedRunningTime="2025-10-13 09:18:10.214275893 +0000 UTC m=+2015.362151674" watchObservedRunningTime="2025-10-13 09:18:10.220493152 +0000 UTC m=+2015.368368903" Oct 13 09:18:12 crc kubenswrapper[4685]: I1013 09:18:12.503125 4685 scope.go:117] "RemoveContainer" containerID="4e57efd2a5e99def1a73daa5f8404b15401378bfed57465d070b35fe0a552d79" Oct 13 09:18:12 crc kubenswrapper[4685]: E1013 09:18:12.505095 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:18:13 crc kubenswrapper[4685]: I1013 09:18:13.595326 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fv4xw" Oct 13 09:18:13 crc kubenswrapper[4685]: I1013 09:18:13.597825 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fv4xw" Oct 13 09:18:13 crc kubenswrapper[4685]: I1013 09:18:13.659616 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fv4xw" Oct 13 09:18:14 crc kubenswrapper[4685]: I1013 09:18:14.281763 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fv4xw" Oct 13 09:18:14 crc kubenswrapper[4685]: I1013 09:18:14.444398 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fv4xw"] Oct 13 09:18:16 crc kubenswrapper[4685]: I1013 09:18:16.252881 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fv4xw" podUID="810071be-34b5-4629-b47a-17a3bb0533dd" containerName="registry-server" containerID="cri-o://a45b788714e9ebd35f78c05f6a2b6447098e9ecac6b956e62dbafdbe0cd32071" gracePeriod=2 Oct 13 09:18:16 crc kubenswrapper[4685]: I1013 09:18:16.722723 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fv4xw" Oct 13 09:18:16 crc kubenswrapper[4685]: I1013 09:18:16.857954 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqfv7\" (UniqueName: \"kubernetes.io/projected/810071be-34b5-4629-b47a-17a3bb0533dd-kube-api-access-jqfv7\") pod \"810071be-34b5-4629-b47a-17a3bb0533dd\" (UID: \"810071be-34b5-4629-b47a-17a3bb0533dd\") " Oct 13 09:18:16 crc kubenswrapper[4685]: I1013 09:18:16.858350 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/810071be-34b5-4629-b47a-17a3bb0533dd-utilities\") pod \"810071be-34b5-4629-b47a-17a3bb0533dd\" (UID: \"810071be-34b5-4629-b47a-17a3bb0533dd\") " Oct 13 09:18:16 crc kubenswrapper[4685]: I1013 09:18:16.859160 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/810071be-34b5-4629-b47a-17a3bb0533dd-utilities" (OuterVolumeSpecName: "utilities") pod "810071be-34b5-4629-b47a-17a3bb0533dd" (UID: "810071be-34b5-4629-b47a-17a3bb0533dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:18:16 crc kubenswrapper[4685]: I1013 09:18:16.859303 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/810071be-34b5-4629-b47a-17a3bb0533dd-catalog-content\") pod \"810071be-34b5-4629-b47a-17a3bb0533dd\" (UID: \"810071be-34b5-4629-b47a-17a3bb0533dd\") " Oct 13 09:18:16 crc kubenswrapper[4685]: I1013 09:18:16.864025 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/810071be-34b5-4629-b47a-17a3bb0533dd-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 09:18:16 crc kubenswrapper[4685]: I1013 09:18:16.864732 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/810071be-34b5-4629-b47a-17a3bb0533dd-kube-api-access-jqfv7" (OuterVolumeSpecName: "kube-api-access-jqfv7") pod "810071be-34b5-4629-b47a-17a3bb0533dd" (UID: "810071be-34b5-4629-b47a-17a3bb0533dd"). InnerVolumeSpecName "kube-api-access-jqfv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:18:16 crc kubenswrapper[4685]: I1013 09:18:16.915316 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/810071be-34b5-4629-b47a-17a3bb0533dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "810071be-34b5-4629-b47a-17a3bb0533dd" (UID: "810071be-34b5-4629-b47a-17a3bb0533dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:18:16 crc kubenswrapper[4685]: I1013 09:18:16.966014 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqfv7\" (UniqueName: \"kubernetes.io/projected/810071be-34b5-4629-b47a-17a3bb0533dd-kube-api-access-jqfv7\") on node \"crc\" DevicePath \"\"" Oct 13 09:18:16 crc kubenswrapper[4685]: I1013 09:18:16.966053 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/810071be-34b5-4629-b47a-17a3bb0533dd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 09:18:17 crc kubenswrapper[4685]: I1013 09:18:17.278816 4685 generic.go:334] "Generic (PLEG): container finished" podID="810071be-34b5-4629-b47a-17a3bb0533dd" containerID="a45b788714e9ebd35f78c05f6a2b6447098e9ecac6b956e62dbafdbe0cd32071" exitCode=0 Oct 13 09:18:17 crc kubenswrapper[4685]: I1013 09:18:17.278870 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv4xw" event={"ID":"810071be-34b5-4629-b47a-17a3bb0533dd","Type":"ContainerDied","Data":"a45b788714e9ebd35f78c05f6a2b6447098e9ecac6b956e62dbafdbe0cd32071"} Oct 13 09:18:17 crc kubenswrapper[4685]: I1013 09:18:17.278949 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv4xw" event={"ID":"810071be-34b5-4629-b47a-17a3bb0533dd","Type":"ContainerDied","Data":"81d37ace3d759d0c7ca41fd9dbfeff3eb4a39f786f6d38f67a9b696906c67e1b"} Oct 13 09:18:17 crc kubenswrapper[4685]: I1013 09:18:17.278985 4685 scope.go:117] "RemoveContainer" containerID="a45b788714e9ebd35f78c05f6a2b6447098e9ecac6b956e62dbafdbe0cd32071" Oct 13 09:18:17 crc kubenswrapper[4685]: I1013 09:18:17.279012 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fv4xw" Oct 13 09:18:17 crc kubenswrapper[4685]: I1013 09:18:17.318021 4685 scope.go:117] "RemoveContainer" containerID="a08c3032469671c4432e0cbeecde7ed2eb22523e011e5bc421c039c34a26c622" Oct 13 09:18:17 crc kubenswrapper[4685]: I1013 09:18:17.324230 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fv4xw"] Oct 13 09:18:17 crc kubenswrapper[4685]: I1013 09:18:17.335655 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fv4xw"] Oct 13 09:18:17 crc kubenswrapper[4685]: I1013 09:18:17.342522 4685 scope.go:117] "RemoveContainer" containerID="75d572c74f7d65e4650127afc71c4b9f87768c8a1b8f8655af95010832aac900" Oct 13 09:18:17 crc kubenswrapper[4685]: I1013 09:18:17.390209 4685 scope.go:117] "RemoveContainer" containerID="a45b788714e9ebd35f78c05f6a2b6447098e9ecac6b956e62dbafdbe0cd32071" Oct 13 09:18:17 crc kubenswrapper[4685]: E1013 09:18:17.390755 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a45b788714e9ebd35f78c05f6a2b6447098e9ecac6b956e62dbafdbe0cd32071\": container with ID starting with a45b788714e9ebd35f78c05f6a2b6447098e9ecac6b956e62dbafdbe0cd32071 not found: ID does not exist" containerID="a45b788714e9ebd35f78c05f6a2b6447098e9ecac6b956e62dbafdbe0cd32071" Oct 13 09:18:17 crc kubenswrapper[4685]: I1013 09:18:17.391070 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a45b788714e9ebd35f78c05f6a2b6447098e9ecac6b956e62dbafdbe0cd32071"} err="failed to get container status \"a45b788714e9ebd35f78c05f6a2b6447098e9ecac6b956e62dbafdbe0cd32071\": rpc error: code = NotFound desc = could not find container \"a45b788714e9ebd35f78c05f6a2b6447098e9ecac6b956e62dbafdbe0cd32071\": container with ID starting with a45b788714e9ebd35f78c05f6a2b6447098e9ecac6b956e62dbafdbe0cd32071 not found: ID does not exist" Oct 13 09:18:17 crc kubenswrapper[4685]: I1013 09:18:17.391243 4685 scope.go:117] "RemoveContainer" containerID="a08c3032469671c4432e0cbeecde7ed2eb22523e011e5bc421c039c34a26c622" Oct 13 09:18:17 crc kubenswrapper[4685]: E1013 09:18:17.391843 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a08c3032469671c4432e0cbeecde7ed2eb22523e011e5bc421c039c34a26c622\": container with ID starting with a08c3032469671c4432e0cbeecde7ed2eb22523e011e5bc421c039c34a26c622 not found: ID does not exist" containerID="a08c3032469671c4432e0cbeecde7ed2eb22523e011e5bc421c039c34a26c622" Oct 13 09:18:17 crc kubenswrapper[4685]: I1013 09:18:17.391890 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a08c3032469671c4432e0cbeecde7ed2eb22523e011e5bc421c039c34a26c622"} err="failed to get container status \"a08c3032469671c4432e0cbeecde7ed2eb22523e011e5bc421c039c34a26c622\": rpc error: code = NotFound desc = could not find container \"a08c3032469671c4432e0cbeecde7ed2eb22523e011e5bc421c039c34a26c622\": container with ID starting with a08c3032469671c4432e0cbeecde7ed2eb22523e011e5bc421c039c34a26c622 not found: ID does not exist" Oct 13 09:18:17 crc kubenswrapper[4685]: I1013 09:18:17.391957 4685 scope.go:117] "RemoveContainer" containerID="75d572c74f7d65e4650127afc71c4b9f87768c8a1b8f8655af95010832aac900" Oct 13 09:18:17 crc kubenswrapper[4685]: E1013 09:18:17.393253 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75d572c74f7d65e4650127afc71c4b9f87768c8a1b8f8655af95010832aac900\": container with ID starting with 75d572c74f7d65e4650127afc71c4b9f87768c8a1b8f8655af95010832aac900 not found: ID does not exist" containerID="75d572c74f7d65e4650127afc71c4b9f87768c8a1b8f8655af95010832aac900" Oct 13 09:18:17 crc kubenswrapper[4685]: I1013 09:18:17.393286 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75d572c74f7d65e4650127afc71c4b9f87768c8a1b8f8655af95010832aac900"} err="failed to get container status \"75d572c74f7d65e4650127afc71c4b9f87768c8a1b8f8655af95010832aac900\": rpc error: code = NotFound desc = could not find container \"75d572c74f7d65e4650127afc71c4b9f87768c8a1b8f8655af95010832aac900\": container with ID starting with 75d572c74f7d65e4650127afc71c4b9f87768c8a1b8f8655af95010832aac900 not found: ID does not exist" Oct 13 09:18:17 crc kubenswrapper[4685]: I1013 09:18:17.513434 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="810071be-34b5-4629-b47a-17a3bb0533dd" path="/var/lib/kubelet/pods/810071be-34b5-4629-b47a-17a3bb0533dd/volumes" Oct 13 09:18:24 crc kubenswrapper[4685]: I1013 09:18:24.502643 4685 scope.go:117] "RemoveContainer" containerID="4e57efd2a5e99def1a73daa5f8404b15401378bfed57465d070b35fe0a552d79" Oct 13 09:18:24 crc kubenswrapper[4685]: E1013 09:18:24.503400 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:18:36 crc kubenswrapper[4685]: I1013 09:18:36.502878 4685 scope.go:117] "RemoveContainer" containerID="4e57efd2a5e99def1a73daa5f8404b15401378bfed57465d070b35fe0a552d79" Oct 13 09:18:36 crc kubenswrapper[4685]: E1013 09:18:36.503543 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:18:47 crc kubenswrapper[4685]: I1013 09:18:47.502651 4685 scope.go:117] "RemoveContainer" containerID="4e57efd2a5e99def1a73daa5f8404b15401378bfed57465d070b35fe0a552d79" Oct 13 09:18:47 crc kubenswrapper[4685]: E1013 09:18:47.503498 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:19:00 crc kubenswrapper[4685]: I1013 09:19:00.502331 4685 scope.go:117] "RemoveContainer" containerID="4e57efd2a5e99def1a73daa5f8404b15401378bfed57465d070b35fe0a552d79" Oct 13 09:19:00 crc kubenswrapper[4685]: E1013 09:19:00.503111 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:19:08 crc kubenswrapper[4685]: I1013 09:19:08.733104 4685 generic.go:334] "Generic (PLEG): container finished" podID="57483ac7-dfa4-4538-8605-11b9215432dc" containerID="f22893f68a2d2c3e5c1282b19433866bd1188a93244ae5750ddd82a80e09f5a1" exitCode=0 Oct 13 09:19:08 crc kubenswrapper[4685]: I1013 09:19:08.733331 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" event={"ID":"57483ac7-dfa4-4538-8605-11b9215432dc","Type":"ContainerDied","Data":"f22893f68a2d2c3e5c1282b19433866bd1188a93244ae5750ddd82a80e09f5a1"} Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.257632 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.445330 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57483ac7-dfa4-4538-8605-11b9215432dc-ovn-combined-ca-bundle\") pod \"57483ac7-dfa4-4538-8605-11b9215432dc\" (UID: \"57483ac7-dfa4-4538-8605-11b9215432dc\") " Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.445406 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57483ac7-dfa4-4538-8605-11b9215432dc-inventory\") pod \"57483ac7-dfa4-4538-8605-11b9215432dc\" (UID: \"57483ac7-dfa4-4538-8605-11b9215432dc\") " Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.445580 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pbdx\" (UniqueName: \"kubernetes.io/projected/57483ac7-dfa4-4538-8605-11b9215432dc-kube-api-access-8pbdx\") pod \"57483ac7-dfa4-4538-8605-11b9215432dc\" (UID: \"57483ac7-dfa4-4538-8605-11b9215432dc\") " Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.445689 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/57483ac7-dfa4-4538-8605-11b9215432dc-ovncontroller-config-0\") pod \"57483ac7-dfa4-4538-8605-11b9215432dc\" (UID: \"57483ac7-dfa4-4538-8605-11b9215432dc\") " Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.445811 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57483ac7-dfa4-4538-8605-11b9215432dc-ssh-key\") pod \"57483ac7-dfa4-4538-8605-11b9215432dc\" (UID: \"57483ac7-dfa4-4538-8605-11b9215432dc\") " Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.452530 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57483ac7-dfa4-4538-8605-11b9215432dc-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "57483ac7-dfa4-4538-8605-11b9215432dc" (UID: "57483ac7-dfa4-4538-8605-11b9215432dc"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.453344 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57483ac7-dfa4-4538-8605-11b9215432dc-kube-api-access-8pbdx" (OuterVolumeSpecName: "kube-api-access-8pbdx") pod "57483ac7-dfa4-4538-8605-11b9215432dc" (UID: "57483ac7-dfa4-4538-8605-11b9215432dc"). InnerVolumeSpecName "kube-api-access-8pbdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.474321 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57483ac7-dfa4-4538-8605-11b9215432dc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "57483ac7-dfa4-4538-8605-11b9215432dc" (UID: "57483ac7-dfa4-4538-8605-11b9215432dc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.481566 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57483ac7-dfa4-4538-8605-11b9215432dc-inventory" (OuterVolumeSpecName: "inventory") pod "57483ac7-dfa4-4538-8605-11b9215432dc" (UID: "57483ac7-dfa4-4538-8605-11b9215432dc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.494320 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57483ac7-dfa4-4538-8605-11b9215432dc-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "57483ac7-dfa4-4538-8605-11b9215432dc" (UID: "57483ac7-dfa4-4538-8605-11b9215432dc"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.563060 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57483ac7-dfa4-4538-8605-11b9215432dc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.563107 4685 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57483ac7-dfa4-4538-8605-11b9215432dc-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.563126 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57483ac7-dfa4-4538-8605-11b9215432dc-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.563143 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pbdx\" (UniqueName: \"kubernetes.io/projected/57483ac7-dfa4-4538-8605-11b9215432dc-kube-api-access-8pbdx\") on node \"crc\" DevicePath \"\"" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.563160 4685 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/57483ac7-dfa4-4538-8605-11b9215432dc-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.755213 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" event={"ID":"57483ac7-dfa4-4538-8605-11b9215432dc","Type":"ContainerDied","Data":"360b3550993e3691b78805643b521474ab2683ad9e5ce60af1af80af91ba1e70"} Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.755276 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="360b3550993e3691b78805643b521474ab2683ad9e5ce60af1af80af91ba1e70" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.755671 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nwnvn" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.931961 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5"] Oct 13 09:19:10 crc kubenswrapper[4685]: E1013 09:19:10.932325 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="810071be-34b5-4629-b47a-17a3bb0533dd" containerName="extract-utilities" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.932341 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="810071be-34b5-4629-b47a-17a3bb0533dd" containerName="extract-utilities" Oct 13 09:19:10 crc kubenswrapper[4685]: E1013 09:19:10.932365 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57483ac7-dfa4-4538-8605-11b9215432dc" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.932372 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="57483ac7-dfa4-4538-8605-11b9215432dc" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 13 09:19:10 crc kubenswrapper[4685]: E1013 09:19:10.932383 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed00208c-afc6-4bc6-89b4-d23b93969110" containerName="registry-server" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.932389 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed00208c-afc6-4bc6-89b4-d23b93969110" containerName="registry-server" Oct 13 09:19:10 crc kubenswrapper[4685]: E1013 09:19:10.932402 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="810071be-34b5-4629-b47a-17a3bb0533dd" containerName="registry-server" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.932408 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="810071be-34b5-4629-b47a-17a3bb0533dd" containerName="registry-server" Oct 13 09:19:10 crc kubenswrapper[4685]: E1013 09:19:10.932424 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed00208c-afc6-4bc6-89b4-d23b93969110" containerName="extract-content" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.932430 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed00208c-afc6-4bc6-89b4-d23b93969110" containerName="extract-content" Oct 13 09:19:10 crc kubenswrapper[4685]: E1013 09:19:10.932451 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed00208c-afc6-4bc6-89b4-d23b93969110" containerName="extract-utilities" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.932457 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed00208c-afc6-4bc6-89b4-d23b93969110" containerName="extract-utilities" Oct 13 09:19:10 crc kubenswrapper[4685]: E1013 09:19:10.932474 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="810071be-34b5-4629-b47a-17a3bb0533dd" containerName="extract-content" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.932480 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="810071be-34b5-4629-b47a-17a3bb0533dd" containerName="extract-content" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.932643 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="810071be-34b5-4629-b47a-17a3bb0533dd" containerName="registry-server" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.932655 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed00208c-afc6-4bc6-89b4-d23b93969110" containerName="registry-server" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.932678 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="57483ac7-dfa4-4538-8605-11b9215432dc" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.933324 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.936298 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.936675 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.936726 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.936784 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.936956 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hcn4z" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.937424 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 13 09:19:10 crc kubenswrapper[4685]: I1013 09:19:10.952342 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5"] Oct 13 09:19:11 crc kubenswrapper[4685]: I1013 09:19:11.070517 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" Oct 13 09:19:11 crc kubenswrapper[4685]: I1013 09:19:11.070614 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" Oct 13 09:19:11 crc kubenswrapper[4685]: I1013 09:19:11.070641 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flx5m\" (UniqueName: \"kubernetes.io/projected/0b2a22b2-f314-42af-95f0-f6213098f535-kube-api-access-flx5m\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" Oct 13 09:19:11 crc kubenswrapper[4685]: I1013 09:19:11.070683 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" Oct 13 09:19:11 crc kubenswrapper[4685]: I1013 09:19:11.070716 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" Oct 13 09:19:11 crc kubenswrapper[4685]: I1013 09:19:11.070746 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" Oct 13 09:19:11 crc kubenswrapper[4685]: I1013 09:19:11.172482 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" Oct 13 09:19:11 crc kubenswrapper[4685]: I1013 09:19:11.172588 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" Oct 13 09:19:11 crc kubenswrapper[4685]: I1013 09:19:11.172617 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flx5m\" (UniqueName: \"kubernetes.io/projected/0b2a22b2-f314-42af-95f0-f6213098f535-kube-api-access-flx5m\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" Oct 13 09:19:11 crc kubenswrapper[4685]: I1013 09:19:11.172686 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" Oct 13 09:19:11 crc kubenswrapper[4685]: I1013 09:19:11.172721 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" Oct 13 09:19:11 crc kubenswrapper[4685]: I1013 09:19:11.172756 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" Oct 13 09:19:11 crc kubenswrapper[4685]: I1013 09:19:11.178139 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" Oct 13 09:19:11 crc kubenswrapper[4685]: I1013 09:19:11.178534 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" Oct 13 09:19:11 crc kubenswrapper[4685]: I1013 09:19:11.179075 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" Oct 13 09:19:11 crc kubenswrapper[4685]: I1013 09:19:11.179144 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" Oct 13 09:19:11 crc kubenswrapper[4685]: I1013 09:19:11.187779 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" Oct 13 09:19:11 crc kubenswrapper[4685]: I1013 09:19:11.192836 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flx5m\" (UniqueName: \"kubernetes.io/projected/0b2a22b2-f314-42af-95f0-f6213098f535-kube-api-access-flx5m\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" Oct 13 09:19:11 crc kubenswrapper[4685]: I1013 09:19:11.253018 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" Oct 13 09:19:11 crc kubenswrapper[4685]: I1013 09:19:11.798165 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5"] Oct 13 09:19:12 crc kubenswrapper[4685]: I1013 09:19:12.772234 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" event={"ID":"0b2a22b2-f314-42af-95f0-f6213098f535","Type":"ContainerStarted","Data":"dda50f147625e6fc49d2667d1b026d740c5ef468267b6342079470222f32e440"} Oct 13 09:19:12 crc kubenswrapper[4685]: I1013 09:19:12.772581 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" event={"ID":"0b2a22b2-f314-42af-95f0-f6213098f535","Type":"ContainerStarted","Data":"eb87afd2d93d5b016bbc6da9dc090c7e79f4e96b4acbd6991e2b56f9488bbbe3"} Oct 13 09:19:12 crc kubenswrapper[4685]: I1013 09:19:12.923446 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" podStartSLOduration=2.420946206 podStartE2EDuration="2.923424963s" podCreationTimestamp="2025-10-13 09:19:10 +0000 UTC" firstStartedPulling="2025-10-13 09:19:11.791585642 +0000 UTC m=+2076.939461423" lastFinishedPulling="2025-10-13 09:19:12.294064419 +0000 UTC m=+2077.441940180" observedRunningTime="2025-10-13 09:19:12.795660743 +0000 UTC m=+2077.943536504" watchObservedRunningTime="2025-10-13 09:19:12.923424963 +0000 UTC m=+2078.071300724" Oct 13 09:19:12 crc kubenswrapper[4685]: I1013 09:19:12.925477 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-69p8x"] Oct 13 09:19:12 crc kubenswrapper[4685]: I1013 09:19:12.927750 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-69p8x" Oct 13 09:19:12 crc kubenswrapper[4685]: I1013 09:19:12.950622 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-69p8x"] Oct 13 09:19:13 crc kubenswrapper[4685]: I1013 09:19:13.016007 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be9b16e1-d8f6-446c-89cc-2539585ba0b8-catalog-content\") pod \"redhat-operators-69p8x\" (UID: \"be9b16e1-d8f6-446c-89cc-2539585ba0b8\") " pod="openshift-marketplace/redhat-operators-69p8x" Oct 13 09:19:13 crc kubenswrapper[4685]: I1013 09:19:13.016419 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be9b16e1-d8f6-446c-89cc-2539585ba0b8-utilities\") pod \"redhat-operators-69p8x\" (UID: \"be9b16e1-d8f6-446c-89cc-2539585ba0b8\") " pod="openshift-marketplace/redhat-operators-69p8x" Oct 13 09:19:13 crc kubenswrapper[4685]: I1013 09:19:13.016557 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fljrr\" (UniqueName: \"kubernetes.io/projected/be9b16e1-d8f6-446c-89cc-2539585ba0b8-kube-api-access-fljrr\") pod \"redhat-operators-69p8x\" (UID: \"be9b16e1-d8f6-446c-89cc-2539585ba0b8\") " pod="openshift-marketplace/redhat-operators-69p8x" Oct 13 09:19:13 crc kubenswrapper[4685]: I1013 09:19:13.118221 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fljrr\" (UniqueName: \"kubernetes.io/projected/be9b16e1-d8f6-446c-89cc-2539585ba0b8-kube-api-access-fljrr\") pod \"redhat-operators-69p8x\" (UID: \"be9b16e1-d8f6-446c-89cc-2539585ba0b8\") " pod="openshift-marketplace/redhat-operators-69p8x" Oct 13 09:19:13 crc kubenswrapper[4685]: I1013 09:19:13.118304 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be9b16e1-d8f6-446c-89cc-2539585ba0b8-catalog-content\") pod \"redhat-operators-69p8x\" (UID: \"be9b16e1-d8f6-446c-89cc-2539585ba0b8\") " pod="openshift-marketplace/redhat-operators-69p8x" Oct 13 09:19:13 crc kubenswrapper[4685]: I1013 09:19:13.118365 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be9b16e1-d8f6-446c-89cc-2539585ba0b8-utilities\") pod \"redhat-operators-69p8x\" (UID: \"be9b16e1-d8f6-446c-89cc-2539585ba0b8\") " pod="openshift-marketplace/redhat-operators-69p8x" Oct 13 09:19:13 crc kubenswrapper[4685]: I1013 09:19:13.118960 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be9b16e1-d8f6-446c-89cc-2539585ba0b8-utilities\") pod \"redhat-operators-69p8x\" (UID: \"be9b16e1-d8f6-446c-89cc-2539585ba0b8\") " pod="openshift-marketplace/redhat-operators-69p8x" Oct 13 09:19:13 crc kubenswrapper[4685]: I1013 09:19:13.118970 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be9b16e1-d8f6-446c-89cc-2539585ba0b8-catalog-content\") pod \"redhat-operators-69p8x\" (UID: \"be9b16e1-d8f6-446c-89cc-2539585ba0b8\") " pod="openshift-marketplace/redhat-operators-69p8x" Oct 13 09:19:13 crc kubenswrapper[4685]: I1013 09:19:13.137694 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fljrr\" (UniqueName: \"kubernetes.io/projected/be9b16e1-d8f6-446c-89cc-2539585ba0b8-kube-api-access-fljrr\") pod \"redhat-operators-69p8x\" (UID: \"be9b16e1-d8f6-446c-89cc-2539585ba0b8\") " pod="openshift-marketplace/redhat-operators-69p8x" Oct 13 09:19:13 crc kubenswrapper[4685]: I1013 09:19:13.250522 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-69p8x" Oct 13 09:19:13 crc kubenswrapper[4685]: I1013 09:19:13.762514 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-69p8x"] Oct 13 09:19:13 crc kubenswrapper[4685]: W1013 09:19:13.766292 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe9b16e1_d8f6_446c_89cc_2539585ba0b8.slice/crio-a9d5ebc68ebe6723a5706eab2ae794951a131598bd1362eebeab8bb4597ac766 WatchSource:0}: Error finding container a9d5ebc68ebe6723a5706eab2ae794951a131598bd1362eebeab8bb4597ac766: Status 404 returned error can't find the container with id a9d5ebc68ebe6723a5706eab2ae794951a131598bd1362eebeab8bb4597ac766 Oct 13 09:19:13 crc kubenswrapper[4685]: I1013 09:19:13.798892 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69p8x" event={"ID":"be9b16e1-d8f6-446c-89cc-2539585ba0b8","Type":"ContainerStarted","Data":"a9d5ebc68ebe6723a5706eab2ae794951a131598bd1362eebeab8bb4597ac766"} Oct 13 09:19:14 crc kubenswrapper[4685]: I1013 09:19:14.809328 4685 generic.go:334] "Generic (PLEG): container finished" podID="be9b16e1-d8f6-446c-89cc-2539585ba0b8" containerID="1ae0eeea2b3f19131527737fb66607ac433710f3f3212e98aeffd84099555c03" exitCode=0 Oct 13 09:19:14 crc kubenswrapper[4685]: I1013 09:19:14.809421 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69p8x" event={"ID":"be9b16e1-d8f6-446c-89cc-2539585ba0b8","Type":"ContainerDied","Data":"1ae0eeea2b3f19131527737fb66607ac433710f3f3212e98aeffd84099555c03"} Oct 13 09:19:15 crc kubenswrapper[4685]: I1013 09:19:15.514816 4685 scope.go:117] "RemoveContainer" containerID="4e57efd2a5e99def1a73daa5f8404b15401378bfed57465d070b35fe0a552d79" Oct 13 09:19:15 crc kubenswrapper[4685]: E1013 09:19:15.517017 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:19:15 crc kubenswrapper[4685]: I1013 09:19:15.822895 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69p8x" event={"ID":"be9b16e1-d8f6-446c-89cc-2539585ba0b8","Type":"ContainerStarted","Data":"e20916c9aeb9b4a85328b0c2d7df644a2fea85b41694d9a412eb8249f6f74372"} Oct 13 09:19:19 crc kubenswrapper[4685]: I1013 09:19:19.864024 4685 generic.go:334] "Generic (PLEG): container finished" podID="be9b16e1-d8f6-446c-89cc-2539585ba0b8" containerID="e20916c9aeb9b4a85328b0c2d7df644a2fea85b41694d9a412eb8249f6f74372" exitCode=0 Oct 13 09:19:19 crc kubenswrapper[4685]: I1013 09:19:19.864296 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69p8x" event={"ID":"be9b16e1-d8f6-446c-89cc-2539585ba0b8","Type":"ContainerDied","Data":"e20916c9aeb9b4a85328b0c2d7df644a2fea85b41694d9a412eb8249f6f74372"} Oct 13 09:19:20 crc kubenswrapper[4685]: I1013 09:19:20.877352 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69p8x" event={"ID":"be9b16e1-d8f6-446c-89cc-2539585ba0b8","Type":"ContainerStarted","Data":"2d89f5d76878bb410d5b49b82b45552570fc8b6b8b76053bc7b4164778fb26e5"} Oct 13 09:19:20 crc kubenswrapper[4685]: I1013 09:19:20.898012 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-69p8x" podStartSLOduration=3.377356088 podStartE2EDuration="8.897900746s" podCreationTimestamp="2025-10-13 09:19:12 +0000 UTC" firstStartedPulling="2025-10-13 09:19:14.811719 +0000 UTC m=+2079.959594771" lastFinishedPulling="2025-10-13 09:19:20.332263668 +0000 UTC m=+2085.480139429" observedRunningTime="2025-10-13 09:19:20.897442813 +0000 UTC m=+2086.045318594" watchObservedRunningTime="2025-10-13 09:19:20.897900746 +0000 UTC m=+2086.045776507" Oct 13 09:19:23 crc kubenswrapper[4685]: I1013 09:19:23.251291 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-69p8x" Oct 13 09:19:23 crc kubenswrapper[4685]: I1013 09:19:23.251597 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-69p8x" Oct 13 09:19:24 crc kubenswrapper[4685]: I1013 09:19:24.305653 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-69p8x" podUID="be9b16e1-d8f6-446c-89cc-2539585ba0b8" containerName="registry-server" probeResult="failure" output=< Oct 13 09:19:24 crc kubenswrapper[4685]: timeout: failed to connect service ":50051" within 1s Oct 13 09:19:24 crc kubenswrapper[4685]: > Oct 13 09:19:29 crc kubenswrapper[4685]: I1013 09:19:29.503693 4685 scope.go:117] "RemoveContainer" containerID="4e57efd2a5e99def1a73daa5f8404b15401378bfed57465d070b35fe0a552d79" Oct 13 09:19:29 crc kubenswrapper[4685]: E1013 09:19:29.504386 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:19:34 crc kubenswrapper[4685]: I1013 09:19:34.304772 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-69p8x" podUID="be9b16e1-d8f6-446c-89cc-2539585ba0b8" containerName="registry-server" probeResult="failure" output=< Oct 13 09:19:34 crc kubenswrapper[4685]: timeout: failed to connect service ":50051" within 1s Oct 13 09:19:34 crc kubenswrapper[4685]: > Oct 13 09:19:43 crc kubenswrapper[4685]: I1013 09:19:43.303595 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-69p8x" Oct 13 09:19:43 crc kubenswrapper[4685]: I1013 09:19:43.348625 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-69p8x" Oct 13 09:19:44 crc kubenswrapper[4685]: I1013 09:19:44.131830 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-69p8x"] Oct 13 09:19:44 crc kubenswrapper[4685]: I1013 09:19:44.502597 4685 scope.go:117] "RemoveContainer" containerID="4e57efd2a5e99def1a73daa5f8404b15401378bfed57465d070b35fe0a552d79" Oct 13 09:19:44 crc kubenswrapper[4685]: E1013 09:19:44.502838 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:19:45 crc kubenswrapper[4685]: I1013 09:19:45.130654 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-69p8x" podUID="be9b16e1-d8f6-446c-89cc-2539585ba0b8" containerName="registry-server" containerID="cri-o://2d89f5d76878bb410d5b49b82b45552570fc8b6b8b76053bc7b4164778fb26e5" gracePeriod=2 Oct 13 09:19:45 crc kubenswrapper[4685]: I1013 09:19:45.649002 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-69p8x" Oct 13 09:19:45 crc kubenswrapper[4685]: I1013 09:19:45.781548 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be9b16e1-d8f6-446c-89cc-2539585ba0b8-utilities\") pod \"be9b16e1-d8f6-446c-89cc-2539585ba0b8\" (UID: \"be9b16e1-d8f6-446c-89cc-2539585ba0b8\") " Oct 13 09:19:45 crc kubenswrapper[4685]: I1013 09:19:45.781665 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fljrr\" (UniqueName: \"kubernetes.io/projected/be9b16e1-d8f6-446c-89cc-2539585ba0b8-kube-api-access-fljrr\") pod \"be9b16e1-d8f6-446c-89cc-2539585ba0b8\" (UID: \"be9b16e1-d8f6-446c-89cc-2539585ba0b8\") " Oct 13 09:19:45 crc kubenswrapper[4685]: I1013 09:19:45.781905 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be9b16e1-d8f6-446c-89cc-2539585ba0b8-catalog-content\") pod \"be9b16e1-d8f6-446c-89cc-2539585ba0b8\" (UID: \"be9b16e1-d8f6-446c-89cc-2539585ba0b8\") " Oct 13 09:19:45 crc kubenswrapper[4685]: I1013 09:19:45.782231 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be9b16e1-d8f6-446c-89cc-2539585ba0b8-utilities" (OuterVolumeSpecName: "utilities") pod "be9b16e1-d8f6-446c-89cc-2539585ba0b8" (UID: "be9b16e1-d8f6-446c-89cc-2539585ba0b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:19:45 crc kubenswrapper[4685]: I1013 09:19:45.782527 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be9b16e1-d8f6-446c-89cc-2539585ba0b8-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 09:19:45 crc kubenswrapper[4685]: I1013 09:19:45.789757 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be9b16e1-d8f6-446c-89cc-2539585ba0b8-kube-api-access-fljrr" (OuterVolumeSpecName: "kube-api-access-fljrr") pod "be9b16e1-d8f6-446c-89cc-2539585ba0b8" (UID: "be9b16e1-d8f6-446c-89cc-2539585ba0b8"). InnerVolumeSpecName "kube-api-access-fljrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:19:45 crc kubenswrapper[4685]: I1013 09:19:45.864875 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be9b16e1-d8f6-446c-89cc-2539585ba0b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be9b16e1-d8f6-446c-89cc-2539585ba0b8" (UID: "be9b16e1-d8f6-446c-89cc-2539585ba0b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:19:45 crc kubenswrapper[4685]: I1013 09:19:45.884647 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fljrr\" (UniqueName: \"kubernetes.io/projected/be9b16e1-d8f6-446c-89cc-2539585ba0b8-kube-api-access-fljrr\") on node \"crc\" DevicePath \"\"" Oct 13 09:19:45 crc kubenswrapper[4685]: I1013 09:19:45.884704 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be9b16e1-d8f6-446c-89cc-2539585ba0b8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 09:19:46 crc kubenswrapper[4685]: I1013 09:19:46.148148 4685 generic.go:334] "Generic (PLEG): container finished" podID="be9b16e1-d8f6-446c-89cc-2539585ba0b8" containerID="2d89f5d76878bb410d5b49b82b45552570fc8b6b8b76053bc7b4164778fb26e5" exitCode=0 Oct 13 09:19:46 crc kubenswrapper[4685]: I1013 09:19:46.148192 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69p8x" event={"ID":"be9b16e1-d8f6-446c-89cc-2539585ba0b8","Type":"ContainerDied","Data":"2d89f5d76878bb410d5b49b82b45552570fc8b6b8b76053bc7b4164778fb26e5"} Oct 13 09:19:46 crc kubenswrapper[4685]: I1013 09:19:46.148230 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69p8x" event={"ID":"be9b16e1-d8f6-446c-89cc-2539585ba0b8","Type":"ContainerDied","Data":"a9d5ebc68ebe6723a5706eab2ae794951a131598bd1362eebeab8bb4597ac766"} Oct 13 09:19:46 crc kubenswrapper[4685]: I1013 09:19:46.148249 4685 scope.go:117] "RemoveContainer" containerID="2d89f5d76878bb410d5b49b82b45552570fc8b6b8b76053bc7b4164778fb26e5" Oct 13 09:19:46 crc kubenswrapper[4685]: I1013 09:19:46.148408 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-69p8x" Oct 13 09:19:46 crc kubenswrapper[4685]: I1013 09:19:46.175988 4685 scope.go:117] "RemoveContainer" containerID="e20916c9aeb9b4a85328b0c2d7df644a2fea85b41694d9a412eb8249f6f74372" Oct 13 09:19:46 crc kubenswrapper[4685]: I1013 09:19:46.205611 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-69p8x"] Oct 13 09:19:46 crc kubenswrapper[4685]: I1013 09:19:46.215198 4685 scope.go:117] "RemoveContainer" containerID="1ae0eeea2b3f19131527737fb66607ac433710f3f3212e98aeffd84099555c03" Oct 13 09:19:46 crc kubenswrapper[4685]: I1013 09:19:46.215211 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-69p8x"] Oct 13 09:19:46 crc kubenswrapper[4685]: I1013 09:19:46.265767 4685 scope.go:117] "RemoveContainer" containerID="2d89f5d76878bb410d5b49b82b45552570fc8b6b8b76053bc7b4164778fb26e5" Oct 13 09:19:46 crc kubenswrapper[4685]: E1013 09:19:46.266377 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d89f5d76878bb410d5b49b82b45552570fc8b6b8b76053bc7b4164778fb26e5\": container with ID starting with 2d89f5d76878bb410d5b49b82b45552570fc8b6b8b76053bc7b4164778fb26e5 not found: ID does not exist" containerID="2d89f5d76878bb410d5b49b82b45552570fc8b6b8b76053bc7b4164778fb26e5" Oct 13 09:19:46 crc kubenswrapper[4685]: I1013 09:19:46.266417 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d89f5d76878bb410d5b49b82b45552570fc8b6b8b76053bc7b4164778fb26e5"} err="failed to get container status \"2d89f5d76878bb410d5b49b82b45552570fc8b6b8b76053bc7b4164778fb26e5\": rpc error: code = NotFound desc = could not find container \"2d89f5d76878bb410d5b49b82b45552570fc8b6b8b76053bc7b4164778fb26e5\": container with ID starting with 2d89f5d76878bb410d5b49b82b45552570fc8b6b8b76053bc7b4164778fb26e5 not found: ID does not exist" Oct 13 09:19:46 crc kubenswrapper[4685]: I1013 09:19:46.266445 4685 scope.go:117] "RemoveContainer" containerID="e20916c9aeb9b4a85328b0c2d7df644a2fea85b41694d9a412eb8249f6f74372" Oct 13 09:19:46 crc kubenswrapper[4685]: E1013 09:19:46.266802 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e20916c9aeb9b4a85328b0c2d7df644a2fea85b41694d9a412eb8249f6f74372\": container with ID starting with e20916c9aeb9b4a85328b0c2d7df644a2fea85b41694d9a412eb8249f6f74372 not found: ID does not exist" containerID="e20916c9aeb9b4a85328b0c2d7df644a2fea85b41694d9a412eb8249f6f74372" Oct 13 09:19:46 crc kubenswrapper[4685]: I1013 09:19:46.266905 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e20916c9aeb9b4a85328b0c2d7df644a2fea85b41694d9a412eb8249f6f74372"} err="failed to get container status \"e20916c9aeb9b4a85328b0c2d7df644a2fea85b41694d9a412eb8249f6f74372\": rpc error: code = NotFound desc = could not find container \"e20916c9aeb9b4a85328b0c2d7df644a2fea85b41694d9a412eb8249f6f74372\": container with ID starting with e20916c9aeb9b4a85328b0c2d7df644a2fea85b41694d9a412eb8249f6f74372 not found: ID does not exist" Oct 13 09:19:46 crc kubenswrapper[4685]: I1013 09:19:46.267019 4685 scope.go:117] "RemoveContainer" containerID="1ae0eeea2b3f19131527737fb66607ac433710f3f3212e98aeffd84099555c03" Oct 13 09:19:46 crc kubenswrapper[4685]: E1013 09:19:46.267506 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ae0eeea2b3f19131527737fb66607ac433710f3f3212e98aeffd84099555c03\": container with ID starting with 1ae0eeea2b3f19131527737fb66607ac433710f3f3212e98aeffd84099555c03 not found: ID does not exist" containerID="1ae0eeea2b3f19131527737fb66607ac433710f3f3212e98aeffd84099555c03" Oct 13 09:19:46 crc kubenswrapper[4685]: I1013 09:19:46.267541 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ae0eeea2b3f19131527737fb66607ac433710f3f3212e98aeffd84099555c03"} err="failed to get container status \"1ae0eeea2b3f19131527737fb66607ac433710f3f3212e98aeffd84099555c03\": rpc error: code = NotFound desc = could not find container \"1ae0eeea2b3f19131527737fb66607ac433710f3f3212e98aeffd84099555c03\": container with ID starting with 1ae0eeea2b3f19131527737fb66607ac433710f3f3212e98aeffd84099555c03 not found: ID does not exist" Oct 13 09:19:47 crc kubenswrapper[4685]: I1013 09:19:47.518640 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be9b16e1-d8f6-446c-89cc-2539585ba0b8" path="/var/lib/kubelet/pods/be9b16e1-d8f6-446c-89cc-2539585ba0b8/volumes" Oct 13 09:19:55 crc kubenswrapper[4685]: I1013 09:19:55.513989 4685 scope.go:117] "RemoveContainer" containerID="4e57efd2a5e99def1a73daa5f8404b15401378bfed57465d070b35fe0a552d79" Oct 13 09:19:56 crc kubenswrapper[4685]: I1013 09:19:56.235971 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" event={"ID":"b001d17a-1aea-44ba-86c5-ba6b312156c1","Type":"ContainerStarted","Data":"6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e"} Oct 13 09:19:56 crc kubenswrapper[4685]: I1013 09:19:56.236687 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:20:03 crc kubenswrapper[4685]: I1013 09:20:03.447719 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:20:04 crc kubenswrapper[4685]: I1013 09:20:04.323189 4685 generic.go:334] "Generic (PLEG): container finished" podID="0b2a22b2-f314-42af-95f0-f6213098f535" containerID="dda50f147625e6fc49d2667d1b026d740c5ef468267b6342079470222f32e440" exitCode=0 Oct 13 09:20:04 crc kubenswrapper[4685]: I1013 09:20:04.323258 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" event={"ID":"0b2a22b2-f314-42af-95f0-f6213098f535","Type":"ContainerDied","Data":"dda50f147625e6fc49d2667d1b026d740c5ef468267b6342079470222f32e440"} Oct 13 09:20:05 crc kubenswrapper[4685]: I1013 09:20:05.748826 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" Oct 13 09:20:05 crc kubenswrapper[4685]: I1013 09:20:05.812291 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-inventory\") pod \"0b2a22b2-f314-42af-95f0-f6213098f535\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " Oct 13 09:20:05 crc kubenswrapper[4685]: I1013 09:20:05.812384 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-neutron-ovn-metadata-agent-neutron-config-0\") pod \"0b2a22b2-f314-42af-95f0-f6213098f535\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " Oct 13 09:20:05 crc kubenswrapper[4685]: I1013 09:20:05.812464 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flx5m\" (UniqueName: \"kubernetes.io/projected/0b2a22b2-f314-42af-95f0-f6213098f535-kube-api-access-flx5m\") pod \"0b2a22b2-f314-42af-95f0-f6213098f535\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " Oct 13 09:20:05 crc kubenswrapper[4685]: I1013 09:20:05.812490 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-nova-metadata-neutron-config-0\") pod \"0b2a22b2-f314-42af-95f0-f6213098f535\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " Oct 13 09:20:05 crc kubenswrapper[4685]: I1013 09:20:05.812537 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-ssh-key\") pod \"0b2a22b2-f314-42af-95f0-f6213098f535\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " Oct 13 09:20:05 crc kubenswrapper[4685]: I1013 09:20:05.812602 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-neutron-metadata-combined-ca-bundle\") pod \"0b2a22b2-f314-42af-95f0-f6213098f535\" (UID: \"0b2a22b2-f314-42af-95f0-f6213098f535\") " Oct 13 09:20:05 crc kubenswrapper[4685]: I1013 09:20:05.818309 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b2a22b2-f314-42af-95f0-f6213098f535-kube-api-access-flx5m" (OuterVolumeSpecName: "kube-api-access-flx5m") pod "0b2a22b2-f314-42af-95f0-f6213098f535" (UID: "0b2a22b2-f314-42af-95f0-f6213098f535"). InnerVolumeSpecName "kube-api-access-flx5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:20:05 crc kubenswrapper[4685]: I1013 09:20:05.818738 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "0b2a22b2-f314-42af-95f0-f6213098f535" (UID: "0b2a22b2-f314-42af-95f0-f6213098f535"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:20:05 crc kubenswrapper[4685]: I1013 09:20:05.846341 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "0b2a22b2-f314-42af-95f0-f6213098f535" (UID: "0b2a22b2-f314-42af-95f0-f6213098f535"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:20:05 crc kubenswrapper[4685]: I1013 09:20:05.848606 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-inventory" (OuterVolumeSpecName: "inventory") pod "0b2a22b2-f314-42af-95f0-f6213098f535" (UID: "0b2a22b2-f314-42af-95f0-f6213098f535"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:20:05 crc kubenswrapper[4685]: I1013 09:20:05.851709 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "0b2a22b2-f314-42af-95f0-f6213098f535" (UID: "0b2a22b2-f314-42af-95f0-f6213098f535"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:20:05 crc kubenswrapper[4685]: I1013 09:20:05.855150 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0b2a22b2-f314-42af-95f0-f6213098f535" (UID: "0b2a22b2-f314-42af-95f0-f6213098f535"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:20:05 crc kubenswrapper[4685]: I1013 09:20:05.914854 4685 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:20:05 crc kubenswrapper[4685]: I1013 09:20:05.914896 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 09:20:05 crc kubenswrapper[4685]: I1013 09:20:05.914908 4685 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 13 09:20:05 crc kubenswrapper[4685]: I1013 09:20:05.914935 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flx5m\" (UniqueName: \"kubernetes.io/projected/0b2a22b2-f314-42af-95f0-f6213098f535-kube-api-access-flx5m\") on node \"crc\" DevicePath \"\"" Oct 13 09:20:05 crc kubenswrapper[4685]: I1013 09:20:05.914945 4685 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 13 09:20:05 crc kubenswrapper[4685]: I1013 09:20:05.914954 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b2a22b2-f314-42af-95f0-f6213098f535-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.338408 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" event={"ID":"0b2a22b2-f314-42af-95f0-f6213098f535","Type":"ContainerDied","Data":"eb87afd2d93d5b016bbc6da9dc090c7e79f4e96b4acbd6991e2b56f9488bbbe3"} Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.338442 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb87afd2d93d5b016bbc6da9dc090c7e79f4e96b4acbd6991e2b56f9488bbbe3" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.338487 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.447125 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2"] Oct 13 09:20:06 crc kubenswrapper[4685]: E1013 09:20:06.447587 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b2a22b2-f314-42af-95f0-f6213098f535" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.447612 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b2a22b2-f314-42af-95f0-f6213098f535" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 13 09:20:06 crc kubenswrapper[4685]: E1013 09:20:06.447653 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be9b16e1-d8f6-446c-89cc-2539585ba0b8" containerName="extract-content" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.447663 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="be9b16e1-d8f6-446c-89cc-2539585ba0b8" containerName="extract-content" Oct 13 09:20:06 crc kubenswrapper[4685]: E1013 09:20:06.447681 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be9b16e1-d8f6-446c-89cc-2539585ba0b8" containerName="extract-utilities" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.447690 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="be9b16e1-d8f6-446c-89cc-2539585ba0b8" containerName="extract-utilities" Oct 13 09:20:06 crc kubenswrapper[4685]: E1013 09:20:06.447729 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be9b16e1-d8f6-446c-89cc-2539585ba0b8" containerName="registry-server" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.447737 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="be9b16e1-d8f6-446c-89cc-2539585ba0b8" containerName="registry-server" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.447991 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b2a22b2-f314-42af-95f0-f6213098f535" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.448037 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="be9b16e1-d8f6-446c-89cc-2539585ba0b8" containerName="registry-server" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.452369 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.454370 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.454376 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.454959 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hcn4z" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.456221 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.459626 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.467940 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2"] Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.525708 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2\" (UID: \"1a2c2127-ae84-45bc-a364-caa9fe26133a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.525855 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2\" (UID: \"1a2c2127-ae84-45bc-a364-caa9fe26133a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.525908 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2\" (UID: \"1a2c2127-ae84-45bc-a364-caa9fe26133a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.526010 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2rgt\" (UniqueName: \"kubernetes.io/projected/1a2c2127-ae84-45bc-a364-caa9fe26133a-kube-api-access-q2rgt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2\" (UID: \"1a2c2127-ae84-45bc-a364-caa9fe26133a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.526068 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2\" (UID: \"1a2c2127-ae84-45bc-a364-caa9fe26133a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.627981 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2\" (UID: \"1a2c2127-ae84-45bc-a364-caa9fe26133a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.628092 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2\" (UID: \"1a2c2127-ae84-45bc-a364-caa9fe26133a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.628224 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2\" (UID: \"1a2c2127-ae84-45bc-a364-caa9fe26133a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.628286 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2\" (UID: \"1a2c2127-ae84-45bc-a364-caa9fe26133a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.628401 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2rgt\" (UniqueName: \"kubernetes.io/projected/1a2c2127-ae84-45bc-a364-caa9fe26133a-kube-api-access-q2rgt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2\" (UID: \"1a2c2127-ae84-45bc-a364-caa9fe26133a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.631948 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2\" (UID: \"1a2c2127-ae84-45bc-a364-caa9fe26133a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.632762 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2\" (UID: \"1a2c2127-ae84-45bc-a364-caa9fe26133a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.638539 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2\" (UID: \"1a2c2127-ae84-45bc-a364-caa9fe26133a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.642750 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2\" (UID: \"1a2c2127-ae84-45bc-a364-caa9fe26133a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.648294 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2rgt\" (UniqueName: \"kubernetes.io/projected/1a2c2127-ae84-45bc-a364-caa9fe26133a-kube-api-access-q2rgt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2\" (UID: \"1a2c2127-ae84-45bc-a364-caa9fe26133a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" Oct 13 09:20:06 crc kubenswrapper[4685]: I1013 09:20:06.770827 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" Oct 13 09:20:07 crc kubenswrapper[4685]: I1013 09:20:07.318894 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2"] Oct 13 09:20:07 crc kubenswrapper[4685]: I1013 09:20:07.347624 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" event={"ID":"1a2c2127-ae84-45bc-a364-caa9fe26133a","Type":"ContainerStarted","Data":"74c4166222a851cfaa985efbba99aeda450634d5d1fb51c31c9ddd07e85f391f"} Oct 13 09:20:08 crc kubenswrapper[4685]: I1013 09:20:08.357093 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" event={"ID":"1a2c2127-ae84-45bc-a364-caa9fe26133a","Type":"ContainerStarted","Data":"0a1832f119690cf74b9271c860fc34dc5084d3867bedef33a7f0b6eefeec2211"} Oct 13 09:20:08 crc kubenswrapper[4685]: I1013 09:20:08.375690 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" podStartSLOduration=1.8875925850000002 podStartE2EDuration="2.375658239s" podCreationTimestamp="2025-10-13 09:20:06 +0000 UTC" firstStartedPulling="2025-10-13 09:20:07.325791921 +0000 UTC m=+2132.473667702" lastFinishedPulling="2025-10-13 09:20:07.813857595 +0000 UTC m=+2132.961733356" observedRunningTime="2025-10-13 09:20:08.370120728 +0000 UTC m=+2133.517996499" watchObservedRunningTime="2025-10-13 09:20:08.375658239 +0000 UTC m=+2133.523534010" Oct 13 09:20:22 crc kubenswrapper[4685]: I1013 09:20:22.980548 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:20:22 crc kubenswrapper[4685]: I1013 09:20:22.981299 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:20:52 crc kubenswrapper[4685]: I1013 09:20:52.980189 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:20:52 crc kubenswrapper[4685]: I1013 09:20:52.980677 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:21:22 crc kubenswrapper[4685]: I1013 09:21:22.980126 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:21:22 crc kubenswrapper[4685]: I1013 09:21:22.980648 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:21:22 crc kubenswrapper[4685]: I1013 09:21:22.980707 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 09:21:22 crc kubenswrapper[4685]: I1013 09:21:22.981489 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1c5982a8f152978248be74b1441fee44fe6cefb88272b0a7d3055c30ef13a327"} pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 09:21:22 crc kubenswrapper[4685]: I1013 09:21:22.981565 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" containerID="cri-o://1c5982a8f152978248be74b1441fee44fe6cefb88272b0a7d3055c30ef13a327" gracePeriod=600 Oct 13 09:21:24 crc kubenswrapper[4685]: I1013 09:21:24.062279 4685 generic.go:334] "Generic (PLEG): container finished" podID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerID="1c5982a8f152978248be74b1441fee44fe6cefb88272b0a7d3055c30ef13a327" exitCode=0 Oct 13 09:21:24 crc kubenswrapper[4685]: I1013 09:21:24.062807 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerDied","Data":"1c5982a8f152978248be74b1441fee44fe6cefb88272b0a7d3055c30ef13a327"} Oct 13 09:21:24 crc kubenswrapper[4685]: I1013 09:21:24.062835 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerStarted","Data":"3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad"} Oct 13 09:21:24 crc kubenswrapper[4685]: I1013 09:21:24.062850 4685 scope.go:117] "RemoveContainer" containerID="199b740df6cb2ff2861a62b82e5d91bfc54e47a360b413fbaec392948d409570" Oct 13 09:22:30 crc kubenswrapper[4685]: I1013 09:22:30.766011 4685 generic.go:334] "Generic (PLEG): container finished" podID="b001d17a-1aea-44ba-86c5-ba6b312156c1" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" exitCode=1 Oct 13 09:22:30 crc kubenswrapper[4685]: I1013 09:22:30.766694 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" event={"ID":"b001d17a-1aea-44ba-86c5-ba6b312156c1","Type":"ContainerDied","Data":"6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e"} Oct 13 09:22:30 crc kubenswrapper[4685]: I1013 09:22:30.766737 4685 scope.go:117] "RemoveContainer" containerID="4e57efd2a5e99def1a73daa5f8404b15401378bfed57465d070b35fe0a552d79" Oct 13 09:22:30 crc kubenswrapper[4685]: I1013 09:22:30.767471 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:22:30 crc kubenswrapper[4685]: E1013 09:22:30.767686 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:22:33 crc kubenswrapper[4685]: I1013 09:22:33.443436 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:22:33 crc kubenswrapper[4685]: I1013 09:22:33.444644 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:22:33 crc kubenswrapper[4685]: E1013 09:22:33.444854 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:22:43 crc kubenswrapper[4685]: I1013 09:22:43.443274 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:22:43 crc kubenswrapper[4685]: I1013 09:22:43.444726 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:22:43 crc kubenswrapper[4685]: E1013 09:22:43.445384 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:22:51 crc kubenswrapper[4685]: I1013 09:22:51.071729 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xpscg"] Oct 13 09:22:51 crc kubenswrapper[4685]: I1013 09:22:51.088854 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xpscg"] Oct 13 09:22:51 crc kubenswrapper[4685]: I1013 09:22:51.088981 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xpscg" Oct 13 09:22:51 crc kubenswrapper[4685]: I1013 09:22:51.237862 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvllm\" (UniqueName: \"kubernetes.io/projected/d954b4d3-0448-4df4-9518-81312c92631c-kube-api-access-hvllm\") pod \"redhat-marketplace-xpscg\" (UID: \"d954b4d3-0448-4df4-9518-81312c92631c\") " pod="openshift-marketplace/redhat-marketplace-xpscg" Oct 13 09:22:51 crc kubenswrapper[4685]: I1013 09:22:51.238680 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d954b4d3-0448-4df4-9518-81312c92631c-catalog-content\") pod \"redhat-marketplace-xpscg\" (UID: \"d954b4d3-0448-4df4-9518-81312c92631c\") " pod="openshift-marketplace/redhat-marketplace-xpscg" Oct 13 09:22:51 crc kubenswrapper[4685]: I1013 09:22:51.238745 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d954b4d3-0448-4df4-9518-81312c92631c-utilities\") pod \"redhat-marketplace-xpscg\" (UID: \"d954b4d3-0448-4df4-9518-81312c92631c\") " pod="openshift-marketplace/redhat-marketplace-xpscg" Oct 13 09:22:51 crc kubenswrapper[4685]: I1013 09:22:51.340181 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvllm\" (UniqueName: \"kubernetes.io/projected/d954b4d3-0448-4df4-9518-81312c92631c-kube-api-access-hvllm\") pod \"redhat-marketplace-xpscg\" (UID: \"d954b4d3-0448-4df4-9518-81312c92631c\") " pod="openshift-marketplace/redhat-marketplace-xpscg" Oct 13 09:22:51 crc kubenswrapper[4685]: I1013 09:22:51.340333 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d954b4d3-0448-4df4-9518-81312c92631c-catalog-content\") pod \"redhat-marketplace-xpscg\" (UID: \"d954b4d3-0448-4df4-9518-81312c92631c\") " pod="openshift-marketplace/redhat-marketplace-xpscg" Oct 13 09:22:51 crc kubenswrapper[4685]: I1013 09:22:51.340374 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d954b4d3-0448-4df4-9518-81312c92631c-utilities\") pod \"redhat-marketplace-xpscg\" (UID: \"d954b4d3-0448-4df4-9518-81312c92631c\") " pod="openshift-marketplace/redhat-marketplace-xpscg" Oct 13 09:22:51 crc kubenswrapper[4685]: I1013 09:22:51.340871 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d954b4d3-0448-4df4-9518-81312c92631c-catalog-content\") pod \"redhat-marketplace-xpscg\" (UID: \"d954b4d3-0448-4df4-9518-81312c92631c\") " pod="openshift-marketplace/redhat-marketplace-xpscg" Oct 13 09:22:51 crc kubenswrapper[4685]: I1013 09:22:51.340940 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d954b4d3-0448-4df4-9518-81312c92631c-utilities\") pod \"redhat-marketplace-xpscg\" (UID: \"d954b4d3-0448-4df4-9518-81312c92631c\") " pod="openshift-marketplace/redhat-marketplace-xpscg" Oct 13 09:22:51 crc kubenswrapper[4685]: I1013 09:22:51.367786 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvllm\" (UniqueName: \"kubernetes.io/projected/d954b4d3-0448-4df4-9518-81312c92631c-kube-api-access-hvllm\") pod \"redhat-marketplace-xpscg\" (UID: \"d954b4d3-0448-4df4-9518-81312c92631c\") " pod="openshift-marketplace/redhat-marketplace-xpscg" Oct 13 09:22:51 crc kubenswrapper[4685]: I1013 09:22:51.455771 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xpscg" Oct 13 09:22:51 crc kubenswrapper[4685]: I1013 09:22:51.837131 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xpscg"] Oct 13 09:22:51 crc kubenswrapper[4685]: I1013 09:22:51.966888 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xpscg" event={"ID":"d954b4d3-0448-4df4-9518-81312c92631c","Type":"ContainerStarted","Data":"e424a7be8ed783216baf22d066959e3051674a67647e3a48c0ce532f61232877"} Oct 13 09:22:52 crc kubenswrapper[4685]: I1013 09:22:52.981645 4685 generic.go:334] "Generic (PLEG): container finished" podID="d954b4d3-0448-4df4-9518-81312c92631c" containerID="f4c021c71c3db1cb3d61f35934470fcf5d71e83d3b409fb0d0e0e28452a24e11" exitCode=0 Oct 13 09:22:52 crc kubenswrapper[4685]: I1013 09:22:52.981722 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xpscg" event={"ID":"d954b4d3-0448-4df4-9518-81312c92631c","Type":"ContainerDied","Data":"f4c021c71c3db1cb3d61f35934470fcf5d71e83d3b409fb0d0e0e28452a24e11"} Oct 13 09:22:52 crc kubenswrapper[4685]: I1013 09:22:52.984551 4685 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 13 09:22:53 crc kubenswrapper[4685]: I1013 09:22:53.991472 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xpscg" event={"ID":"d954b4d3-0448-4df4-9518-81312c92631c","Type":"ContainerStarted","Data":"15f363bd9a83921a10375431f6a0513e383660f3c3cc7c59fe954051d5fdcb23"} Oct 13 09:22:55 crc kubenswrapper[4685]: I1013 09:22:55.003431 4685 generic.go:334] "Generic (PLEG): container finished" podID="d954b4d3-0448-4df4-9518-81312c92631c" containerID="15f363bd9a83921a10375431f6a0513e383660f3c3cc7c59fe954051d5fdcb23" exitCode=0 Oct 13 09:22:55 crc kubenswrapper[4685]: I1013 09:22:55.003490 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xpscg" event={"ID":"d954b4d3-0448-4df4-9518-81312c92631c","Type":"ContainerDied","Data":"15f363bd9a83921a10375431f6a0513e383660f3c3cc7c59fe954051d5fdcb23"} Oct 13 09:22:55 crc kubenswrapper[4685]: I1013 09:22:55.509533 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:22:55 crc kubenswrapper[4685]: E1013 09:22:55.510047 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:22:56 crc kubenswrapper[4685]: I1013 09:22:56.013091 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xpscg" event={"ID":"d954b4d3-0448-4df4-9518-81312c92631c","Type":"ContainerStarted","Data":"0f885268fc3522206f95b1274d852c83b56c260b36891ee330d564ee6c62901c"} Oct 13 09:22:56 crc kubenswrapper[4685]: I1013 09:22:56.033445 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xpscg" podStartSLOduration=2.610526703 podStartE2EDuration="5.033423463s" podCreationTimestamp="2025-10-13 09:22:51 +0000 UTC" firstStartedPulling="2025-10-13 09:22:52.984234077 +0000 UTC m=+2298.132109838" lastFinishedPulling="2025-10-13 09:22:55.407130837 +0000 UTC m=+2300.555006598" observedRunningTime="2025-10-13 09:22:56.030450953 +0000 UTC m=+2301.178326714" watchObservedRunningTime="2025-10-13 09:22:56.033423463 +0000 UTC m=+2301.181299224" Oct 13 09:23:01 crc kubenswrapper[4685]: I1013 09:23:01.456959 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xpscg" Oct 13 09:23:01 crc kubenswrapper[4685]: I1013 09:23:01.457475 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xpscg" Oct 13 09:23:01 crc kubenswrapper[4685]: I1013 09:23:01.517040 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xpscg" Oct 13 09:23:02 crc kubenswrapper[4685]: I1013 09:23:02.159969 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xpscg" Oct 13 09:23:02 crc kubenswrapper[4685]: I1013 09:23:02.243679 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xpscg"] Oct 13 09:23:04 crc kubenswrapper[4685]: I1013 09:23:04.081815 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xpscg" podUID="d954b4d3-0448-4df4-9518-81312c92631c" containerName="registry-server" containerID="cri-o://0f885268fc3522206f95b1274d852c83b56c260b36891ee330d564ee6c62901c" gracePeriod=2 Oct 13 09:23:04 crc kubenswrapper[4685]: I1013 09:23:04.561736 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xpscg" Oct 13 09:23:04 crc kubenswrapper[4685]: I1013 09:23:04.718161 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvllm\" (UniqueName: \"kubernetes.io/projected/d954b4d3-0448-4df4-9518-81312c92631c-kube-api-access-hvllm\") pod \"d954b4d3-0448-4df4-9518-81312c92631c\" (UID: \"d954b4d3-0448-4df4-9518-81312c92631c\") " Oct 13 09:23:04 crc kubenswrapper[4685]: I1013 09:23:04.718263 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d954b4d3-0448-4df4-9518-81312c92631c-catalog-content\") pod \"d954b4d3-0448-4df4-9518-81312c92631c\" (UID: \"d954b4d3-0448-4df4-9518-81312c92631c\") " Oct 13 09:23:04 crc kubenswrapper[4685]: I1013 09:23:04.718496 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d954b4d3-0448-4df4-9518-81312c92631c-utilities\") pod \"d954b4d3-0448-4df4-9518-81312c92631c\" (UID: \"d954b4d3-0448-4df4-9518-81312c92631c\") " Oct 13 09:23:04 crc kubenswrapper[4685]: I1013 09:23:04.719870 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d954b4d3-0448-4df4-9518-81312c92631c-utilities" (OuterVolumeSpecName: "utilities") pod "d954b4d3-0448-4df4-9518-81312c92631c" (UID: "d954b4d3-0448-4df4-9518-81312c92631c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:23:04 crc kubenswrapper[4685]: I1013 09:23:04.727218 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d954b4d3-0448-4df4-9518-81312c92631c-kube-api-access-hvllm" (OuterVolumeSpecName: "kube-api-access-hvllm") pod "d954b4d3-0448-4df4-9518-81312c92631c" (UID: "d954b4d3-0448-4df4-9518-81312c92631c"). InnerVolumeSpecName "kube-api-access-hvllm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:23:04 crc kubenswrapper[4685]: I1013 09:23:04.732471 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d954b4d3-0448-4df4-9518-81312c92631c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d954b4d3-0448-4df4-9518-81312c92631c" (UID: "d954b4d3-0448-4df4-9518-81312c92631c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:23:04 crc kubenswrapper[4685]: I1013 09:23:04.821119 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvllm\" (UniqueName: \"kubernetes.io/projected/d954b4d3-0448-4df4-9518-81312c92631c-kube-api-access-hvllm\") on node \"crc\" DevicePath \"\"" Oct 13 09:23:04 crc kubenswrapper[4685]: I1013 09:23:04.821163 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d954b4d3-0448-4df4-9518-81312c92631c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 09:23:04 crc kubenswrapper[4685]: I1013 09:23:04.821175 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d954b4d3-0448-4df4-9518-81312c92631c-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 09:23:05 crc kubenswrapper[4685]: I1013 09:23:05.093675 4685 generic.go:334] "Generic (PLEG): container finished" podID="d954b4d3-0448-4df4-9518-81312c92631c" containerID="0f885268fc3522206f95b1274d852c83b56c260b36891ee330d564ee6c62901c" exitCode=0 Oct 13 09:23:05 crc kubenswrapper[4685]: I1013 09:23:05.093749 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xpscg" event={"ID":"d954b4d3-0448-4df4-9518-81312c92631c","Type":"ContainerDied","Data":"0f885268fc3522206f95b1274d852c83b56c260b36891ee330d564ee6c62901c"} Oct 13 09:23:05 crc kubenswrapper[4685]: I1013 09:23:05.093761 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xpscg" Oct 13 09:23:05 crc kubenswrapper[4685]: I1013 09:23:05.093813 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xpscg" event={"ID":"d954b4d3-0448-4df4-9518-81312c92631c","Type":"ContainerDied","Data":"e424a7be8ed783216baf22d066959e3051674a67647e3a48c0ce532f61232877"} Oct 13 09:23:05 crc kubenswrapper[4685]: I1013 09:23:05.093841 4685 scope.go:117] "RemoveContainer" containerID="0f885268fc3522206f95b1274d852c83b56c260b36891ee330d564ee6c62901c" Oct 13 09:23:05 crc kubenswrapper[4685]: I1013 09:23:05.126241 4685 scope.go:117] "RemoveContainer" containerID="15f363bd9a83921a10375431f6a0513e383660f3c3cc7c59fe954051d5fdcb23" Oct 13 09:23:05 crc kubenswrapper[4685]: I1013 09:23:05.163868 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xpscg"] Oct 13 09:23:05 crc kubenswrapper[4685]: I1013 09:23:05.172985 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xpscg"] Oct 13 09:23:05 crc kubenswrapper[4685]: I1013 09:23:05.173156 4685 scope.go:117] "RemoveContainer" containerID="f4c021c71c3db1cb3d61f35934470fcf5d71e83d3b409fb0d0e0e28452a24e11" Oct 13 09:23:05 crc kubenswrapper[4685]: E1013 09:23:05.202669 4685 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd954b4d3_0448_4df4_9518_81312c92631c.slice/crio-e424a7be8ed783216baf22d066959e3051674a67647e3a48c0ce532f61232877\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd954b4d3_0448_4df4_9518_81312c92631c.slice\": RecentStats: unable to find data in memory cache]" Oct 13 09:23:05 crc kubenswrapper[4685]: I1013 09:23:05.216199 4685 scope.go:117] "RemoveContainer" containerID="0f885268fc3522206f95b1274d852c83b56c260b36891ee330d564ee6c62901c" Oct 13 09:23:05 crc kubenswrapper[4685]: E1013 09:23:05.216500 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f885268fc3522206f95b1274d852c83b56c260b36891ee330d564ee6c62901c\": container with ID starting with 0f885268fc3522206f95b1274d852c83b56c260b36891ee330d564ee6c62901c not found: ID does not exist" containerID="0f885268fc3522206f95b1274d852c83b56c260b36891ee330d564ee6c62901c" Oct 13 09:23:05 crc kubenswrapper[4685]: I1013 09:23:05.216533 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f885268fc3522206f95b1274d852c83b56c260b36891ee330d564ee6c62901c"} err="failed to get container status \"0f885268fc3522206f95b1274d852c83b56c260b36891ee330d564ee6c62901c\": rpc error: code = NotFound desc = could not find container \"0f885268fc3522206f95b1274d852c83b56c260b36891ee330d564ee6c62901c\": container with ID starting with 0f885268fc3522206f95b1274d852c83b56c260b36891ee330d564ee6c62901c not found: ID does not exist" Oct 13 09:23:05 crc kubenswrapper[4685]: I1013 09:23:05.216558 4685 scope.go:117] "RemoveContainer" containerID="15f363bd9a83921a10375431f6a0513e383660f3c3cc7c59fe954051d5fdcb23" Oct 13 09:23:05 crc kubenswrapper[4685]: E1013 09:23:05.216845 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15f363bd9a83921a10375431f6a0513e383660f3c3cc7c59fe954051d5fdcb23\": container with ID starting with 15f363bd9a83921a10375431f6a0513e383660f3c3cc7c59fe954051d5fdcb23 not found: ID does not exist" containerID="15f363bd9a83921a10375431f6a0513e383660f3c3cc7c59fe954051d5fdcb23" Oct 13 09:23:05 crc kubenswrapper[4685]: I1013 09:23:05.217065 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15f363bd9a83921a10375431f6a0513e383660f3c3cc7c59fe954051d5fdcb23"} err="failed to get container status \"15f363bd9a83921a10375431f6a0513e383660f3c3cc7c59fe954051d5fdcb23\": rpc error: code = NotFound desc = could not find container \"15f363bd9a83921a10375431f6a0513e383660f3c3cc7c59fe954051d5fdcb23\": container with ID starting with 15f363bd9a83921a10375431f6a0513e383660f3c3cc7c59fe954051d5fdcb23 not found: ID does not exist" Oct 13 09:23:05 crc kubenswrapper[4685]: I1013 09:23:05.217203 4685 scope.go:117] "RemoveContainer" containerID="f4c021c71c3db1cb3d61f35934470fcf5d71e83d3b409fb0d0e0e28452a24e11" Oct 13 09:23:05 crc kubenswrapper[4685]: E1013 09:23:05.218074 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4c021c71c3db1cb3d61f35934470fcf5d71e83d3b409fb0d0e0e28452a24e11\": container with ID starting with f4c021c71c3db1cb3d61f35934470fcf5d71e83d3b409fb0d0e0e28452a24e11 not found: ID does not exist" containerID="f4c021c71c3db1cb3d61f35934470fcf5d71e83d3b409fb0d0e0e28452a24e11" Oct 13 09:23:05 crc kubenswrapper[4685]: I1013 09:23:05.218100 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4c021c71c3db1cb3d61f35934470fcf5d71e83d3b409fb0d0e0e28452a24e11"} err="failed to get container status \"f4c021c71c3db1cb3d61f35934470fcf5d71e83d3b409fb0d0e0e28452a24e11\": rpc error: code = NotFound desc = could not find container \"f4c021c71c3db1cb3d61f35934470fcf5d71e83d3b409fb0d0e0e28452a24e11\": container with ID starting with f4c021c71c3db1cb3d61f35934470fcf5d71e83d3b409fb0d0e0e28452a24e11 not found: ID does not exist" Oct 13 09:23:05 crc kubenswrapper[4685]: I1013 09:23:05.529862 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d954b4d3-0448-4df4-9518-81312c92631c" path="/var/lib/kubelet/pods/d954b4d3-0448-4df4-9518-81312c92631c/volumes" Oct 13 09:23:06 crc kubenswrapper[4685]: I1013 09:23:06.503218 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:23:06 crc kubenswrapper[4685]: E1013 09:23:06.503492 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:23:21 crc kubenswrapper[4685]: I1013 09:23:21.503468 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:23:21 crc kubenswrapper[4685]: E1013 09:23:21.504213 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:23:34 crc kubenswrapper[4685]: I1013 09:23:34.502896 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:23:34 crc kubenswrapper[4685]: E1013 09:23:34.503852 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:23:46 crc kubenswrapper[4685]: I1013 09:23:46.503556 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:23:46 crc kubenswrapper[4685]: E1013 09:23:46.504479 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:23:52 crc kubenswrapper[4685]: I1013 09:23:52.980215 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:23:52 crc kubenswrapper[4685]: I1013 09:23:52.980751 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:24:01 crc kubenswrapper[4685]: I1013 09:24:01.502787 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:24:01 crc kubenswrapper[4685]: E1013 09:24:01.503621 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:24:15 crc kubenswrapper[4685]: I1013 09:24:15.514278 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:24:15 crc kubenswrapper[4685]: E1013 09:24:15.515360 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:24:22 crc kubenswrapper[4685]: I1013 09:24:22.979753 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:24:22 crc kubenswrapper[4685]: I1013 09:24:22.980822 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:24:29 crc kubenswrapper[4685]: I1013 09:24:29.504160 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:24:29 crc kubenswrapper[4685]: E1013 09:24:29.507530 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:24:33 crc kubenswrapper[4685]: I1013 09:24:33.907977 4685 generic.go:334] "Generic (PLEG): container finished" podID="1a2c2127-ae84-45bc-a364-caa9fe26133a" containerID="0a1832f119690cf74b9271c860fc34dc5084d3867bedef33a7f0b6eefeec2211" exitCode=0 Oct 13 09:24:33 crc kubenswrapper[4685]: I1013 09:24:33.908039 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" event={"ID":"1a2c2127-ae84-45bc-a364-caa9fe26133a","Type":"ContainerDied","Data":"0a1832f119690cf74b9271c860fc34dc5084d3867bedef33a7f0b6eefeec2211"} Oct 13 09:24:35 crc kubenswrapper[4685]: I1013 09:24:35.318071 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" Oct 13 09:24:35 crc kubenswrapper[4685]: I1013 09:24:35.470013 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-libvirt-secret-0\") pod \"1a2c2127-ae84-45bc-a364-caa9fe26133a\" (UID: \"1a2c2127-ae84-45bc-a364-caa9fe26133a\") " Oct 13 09:24:35 crc kubenswrapper[4685]: I1013 09:24:35.470174 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-inventory\") pod \"1a2c2127-ae84-45bc-a364-caa9fe26133a\" (UID: \"1a2c2127-ae84-45bc-a364-caa9fe26133a\") " Oct 13 09:24:35 crc kubenswrapper[4685]: I1013 09:24:35.470279 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2rgt\" (UniqueName: \"kubernetes.io/projected/1a2c2127-ae84-45bc-a364-caa9fe26133a-kube-api-access-q2rgt\") pod \"1a2c2127-ae84-45bc-a364-caa9fe26133a\" (UID: \"1a2c2127-ae84-45bc-a364-caa9fe26133a\") " Oct 13 09:24:35 crc kubenswrapper[4685]: I1013 09:24:35.470345 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-libvirt-combined-ca-bundle\") pod \"1a2c2127-ae84-45bc-a364-caa9fe26133a\" (UID: \"1a2c2127-ae84-45bc-a364-caa9fe26133a\") " Oct 13 09:24:35 crc kubenswrapper[4685]: I1013 09:24:35.470451 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-ssh-key\") pod \"1a2c2127-ae84-45bc-a364-caa9fe26133a\" (UID: \"1a2c2127-ae84-45bc-a364-caa9fe26133a\") " Oct 13 09:24:35 crc kubenswrapper[4685]: I1013 09:24:35.492004 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "1a2c2127-ae84-45bc-a364-caa9fe26133a" (UID: "1a2c2127-ae84-45bc-a364-caa9fe26133a"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:24:35 crc kubenswrapper[4685]: I1013 09:24:35.504171 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a2c2127-ae84-45bc-a364-caa9fe26133a-kube-api-access-q2rgt" (OuterVolumeSpecName: "kube-api-access-q2rgt") pod "1a2c2127-ae84-45bc-a364-caa9fe26133a" (UID: "1a2c2127-ae84-45bc-a364-caa9fe26133a"). InnerVolumeSpecName "kube-api-access-q2rgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:24:35 crc kubenswrapper[4685]: I1013 09:24:35.535232 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-inventory" (OuterVolumeSpecName: "inventory") pod "1a2c2127-ae84-45bc-a364-caa9fe26133a" (UID: "1a2c2127-ae84-45bc-a364-caa9fe26133a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:24:35 crc kubenswrapper[4685]: I1013 09:24:35.537256 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "1a2c2127-ae84-45bc-a364-caa9fe26133a" (UID: "1a2c2127-ae84-45bc-a364-caa9fe26133a"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:24:35 crc kubenswrapper[4685]: I1013 09:24:35.556439 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1a2c2127-ae84-45bc-a364-caa9fe26133a" (UID: "1a2c2127-ae84-45bc-a364-caa9fe26133a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:24:35 crc kubenswrapper[4685]: I1013 09:24:35.572964 4685 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 13 09:24:35 crc kubenswrapper[4685]: I1013 09:24:35.573004 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 09:24:35 crc kubenswrapper[4685]: I1013 09:24:35.573016 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2rgt\" (UniqueName: \"kubernetes.io/projected/1a2c2127-ae84-45bc-a364-caa9fe26133a-kube-api-access-q2rgt\") on node \"crc\" DevicePath \"\"" Oct 13 09:24:35 crc kubenswrapper[4685]: I1013 09:24:35.573028 4685 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:24:35 crc kubenswrapper[4685]: I1013 09:24:35.573040 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a2c2127-ae84-45bc-a364-caa9fe26133a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 09:24:35 crc kubenswrapper[4685]: I1013 09:24:35.932160 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" event={"ID":"1a2c2127-ae84-45bc-a364-caa9fe26133a","Type":"ContainerDied","Data":"74c4166222a851cfaa985efbba99aeda450634d5d1fb51c31c9ddd07e85f391f"} Oct 13 09:24:35 crc kubenswrapper[4685]: I1013 09:24:35.932219 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74c4166222a851cfaa985efbba99aeda450634d5d1fb51c31c9ddd07e85f391f" Oct 13 09:24:35 crc kubenswrapper[4685]: I1013 09:24:35.932239 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.059354 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw"] Oct 13 09:24:36 crc kubenswrapper[4685]: E1013 09:24:36.059702 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d954b4d3-0448-4df4-9518-81312c92631c" containerName="extract-content" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.059718 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="d954b4d3-0448-4df4-9518-81312c92631c" containerName="extract-content" Oct 13 09:24:36 crc kubenswrapper[4685]: E1013 09:24:36.059732 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d954b4d3-0448-4df4-9518-81312c92631c" containerName="registry-server" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.059738 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="d954b4d3-0448-4df4-9518-81312c92631c" containerName="registry-server" Oct 13 09:24:36 crc kubenswrapper[4685]: E1013 09:24:36.059764 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d954b4d3-0448-4df4-9518-81312c92631c" containerName="extract-utilities" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.059771 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="d954b4d3-0448-4df4-9518-81312c92631c" containerName="extract-utilities" Oct 13 09:24:36 crc kubenswrapper[4685]: E1013 09:24:36.059785 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a2c2127-ae84-45bc-a364-caa9fe26133a" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.059793 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a2c2127-ae84-45bc-a364-caa9fe26133a" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.059974 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a2c2127-ae84-45bc-a364-caa9fe26133a" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.059987 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="d954b4d3-0448-4df4-9518-81312c92631c" containerName="registry-server" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.060613 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.062380 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.064055 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.064275 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.064396 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.064502 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.064617 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.071062 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hcn4z" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.082671 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw"] Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.185970 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.186025 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.186069 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.186087 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.186104 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-796zb\" (UniqueName: \"kubernetes.io/projected/5e5b7b28-ee18-4b85-af9c-030cab7da493-kube-api-access-796zb\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.186246 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.186516 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.186656 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.186689 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.288281 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.288356 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.288408 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.288438 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.288461 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-796zb\" (UniqueName: \"kubernetes.io/projected/5e5b7b28-ee18-4b85-af9c-030cab7da493-kube-api-access-796zb\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.288501 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.288574 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.288620 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.288642 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.290541 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.294391 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.294466 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.296811 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.297318 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.305807 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.308873 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.309776 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.312430 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-796zb\" (UniqueName: \"kubernetes.io/projected/5e5b7b28-ee18-4b85-af9c-030cab7da493-kube-api-access-796zb\") pod \"nova-edpm-deployment-openstack-edpm-ipam-f2kmw\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.396463 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:24:36 crc kubenswrapper[4685]: I1013 09:24:36.952684 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw"] Oct 13 09:24:37 crc kubenswrapper[4685]: I1013 09:24:37.955464 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" event={"ID":"5e5b7b28-ee18-4b85-af9c-030cab7da493","Type":"ContainerStarted","Data":"a081020290e10de2b2de1c5c3feaaeada9bd9acf03dea88a1b9edc39c6260705"} Oct 13 09:24:37 crc kubenswrapper[4685]: I1013 09:24:37.955848 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" event={"ID":"5e5b7b28-ee18-4b85-af9c-030cab7da493","Type":"ContainerStarted","Data":"606a66da1cee419df6eb14979eadb59edf475f42a22b87d2fdb37a6b6f8f7d85"} Oct 13 09:24:38 crc kubenswrapper[4685]: I1013 09:24:38.005154 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" podStartSLOduration=1.510883533 podStartE2EDuration="2.005135054s" podCreationTimestamp="2025-10-13 09:24:36 +0000 UTC" firstStartedPulling="2025-10-13 09:24:36.956353364 +0000 UTC m=+2402.104229125" lastFinishedPulling="2025-10-13 09:24:37.450604885 +0000 UTC m=+2402.598480646" observedRunningTime="2025-10-13 09:24:38.000868339 +0000 UTC m=+2403.148744120" watchObservedRunningTime="2025-10-13 09:24:38.005135054 +0000 UTC m=+2403.153010815" Oct 13 09:24:42 crc kubenswrapper[4685]: I1013 09:24:42.502331 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:24:42 crc kubenswrapper[4685]: E1013 09:24:42.504170 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:24:52 crc kubenswrapper[4685]: I1013 09:24:52.980613 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:24:52 crc kubenswrapper[4685]: I1013 09:24:52.981270 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:24:52 crc kubenswrapper[4685]: I1013 09:24:52.981327 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 09:24:52 crc kubenswrapper[4685]: I1013 09:24:52.982308 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad"} pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 09:24:52 crc kubenswrapper[4685]: I1013 09:24:52.982473 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" containerID="cri-o://3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" gracePeriod=600 Oct 13 09:24:53 crc kubenswrapper[4685]: E1013 09:24:53.116469 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:24:54 crc kubenswrapper[4685]: I1013 09:24:54.104904 4685 generic.go:334] "Generic (PLEG): container finished" podID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" exitCode=0 Oct 13 09:24:54 crc kubenswrapper[4685]: I1013 09:24:54.104962 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerDied","Data":"3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad"} Oct 13 09:24:54 crc kubenswrapper[4685]: I1013 09:24:54.105039 4685 scope.go:117] "RemoveContainer" containerID="1c5982a8f152978248be74b1441fee44fe6cefb88272b0a7d3055c30ef13a327" Oct 13 09:24:54 crc kubenswrapper[4685]: I1013 09:24:54.105966 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:24:54 crc kubenswrapper[4685]: E1013 09:24:54.106302 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:24:55 crc kubenswrapper[4685]: I1013 09:24:55.510280 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:24:55 crc kubenswrapper[4685]: E1013 09:24:55.511024 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:25:05 crc kubenswrapper[4685]: I1013 09:25:05.511107 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:25:05 crc kubenswrapper[4685]: E1013 09:25:05.511892 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:25:06 crc kubenswrapper[4685]: I1013 09:25:06.503152 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:25:06 crc kubenswrapper[4685]: E1013 09:25:06.503427 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:25:17 crc kubenswrapper[4685]: I1013 09:25:17.503207 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:25:17 crc kubenswrapper[4685]: E1013 09:25:17.504102 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:25:19 crc kubenswrapper[4685]: I1013 09:25:19.503324 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:25:19 crc kubenswrapper[4685]: E1013 09:25:19.503978 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:25:28 crc kubenswrapper[4685]: I1013 09:25:28.503273 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:25:28 crc kubenswrapper[4685]: E1013 09:25:28.504245 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:25:33 crc kubenswrapper[4685]: I1013 09:25:33.504225 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:25:33 crc kubenswrapper[4685]: E1013 09:25:33.506620 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:25:41 crc kubenswrapper[4685]: I1013 09:25:41.523231 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:25:41 crc kubenswrapper[4685]: E1013 09:25:41.524421 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:25:44 crc kubenswrapper[4685]: I1013 09:25:44.503426 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:25:44 crc kubenswrapper[4685]: E1013 09:25:44.504609 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:25:53 crc kubenswrapper[4685]: I1013 09:25:53.502971 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:25:53 crc kubenswrapper[4685]: E1013 09:25:53.504308 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:25:58 crc kubenswrapper[4685]: I1013 09:25:58.502359 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:25:58 crc kubenswrapper[4685]: E1013 09:25:58.504650 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:26:08 crc kubenswrapper[4685]: I1013 09:26:08.502607 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:26:08 crc kubenswrapper[4685]: E1013 09:26:08.503395 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:26:09 crc kubenswrapper[4685]: I1013 09:26:09.503981 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:26:09 crc kubenswrapper[4685]: E1013 09:26:09.504282 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:26:20 crc kubenswrapper[4685]: I1013 09:26:20.502802 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:26:20 crc kubenswrapper[4685]: E1013 09:26:20.504064 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:26:22 crc kubenswrapper[4685]: I1013 09:26:22.503939 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:26:22 crc kubenswrapper[4685]: E1013 09:26:22.504582 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:26:31 crc kubenswrapper[4685]: I1013 09:26:31.503138 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:26:31 crc kubenswrapper[4685]: E1013 09:26:31.504586 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:26:37 crc kubenswrapper[4685]: I1013 09:26:37.503201 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:26:37 crc kubenswrapper[4685]: E1013 09:26:37.504701 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:26:44 crc kubenswrapper[4685]: I1013 09:26:44.502750 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:26:44 crc kubenswrapper[4685]: E1013 09:26:44.503621 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:26:50 crc kubenswrapper[4685]: I1013 09:26:50.502863 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:26:50 crc kubenswrapper[4685]: E1013 09:26:50.503713 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:26:58 crc kubenswrapper[4685]: I1013 09:26:58.503093 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:26:58 crc kubenswrapper[4685]: E1013 09:26:58.503802 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:27:05 crc kubenswrapper[4685]: I1013 09:27:05.510836 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:27:05 crc kubenswrapper[4685]: E1013 09:27:05.511775 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:27:09 crc kubenswrapper[4685]: I1013 09:27:09.503567 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:27:09 crc kubenswrapper[4685]: E1013 09:27:09.504213 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:27:17 crc kubenswrapper[4685]: I1013 09:27:17.503806 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:27:17 crc kubenswrapper[4685]: E1013 09:27:17.504533 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:27:23 crc kubenswrapper[4685]: I1013 09:27:23.503413 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:27:23 crc kubenswrapper[4685]: E1013 09:27:23.504345 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:27:31 crc kubenswrapper[4685]: I1013 09:27:31.503388 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:27:32 crc kubenswrapper[4685]: I1013 09:27:32.564239 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" event={"ID":"b001d17a-1aea-44ba-86c5-ba6b312156c1","Type":"ContainerStarted","Data":"69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b"} Oct 13 09:27:32 crc kubenswrapper[4685]: I1013 09:27:32.564980 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:27:34 crc kubenswrapper[4685]: I1013 09:27:34.502897 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:27:34 crc kubenswrapper[4685]: E1013 09:27:34.503185 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:27:43 crc kubenswrapper[4685]: I1013 09:27:43.445369 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:27:47 crc kubenswrapper[4685]: I1013 09:27:47.502872 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:27:47 crc kubenswrapper[4685]: E1013 09:27:47.503660 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:27:59 crc kubenswrapper[4685]: I1013 09:27:59.503591 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:27:59 crc kubenswrapper[4685]: E1013 09:27:59.504313 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:28:06 crc kubenswrapper[4685]: I1013 09:28:06.871136 4685 generic.go:334] "Generic (PLEG): container finished" podID="5e5b7b28-ee18-4b85-af9c-030cab7da493" containerID="a081020290e10de2b2de1c5c3feaaeada9bd9acf03dea88a1b9edc39c6260705" exitCode=0 Oct 13 09:28:06 crc kubenswrapper[4685]: I1013 09:28:06.871227 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" event={"ID":"5e5b7b28-ee18-4b85-af9c-030cab7da493","Type":"ContainerDied","Data":"a081020290e10de2b2de1c5c3feaaeada9bd9acf03dea88a1b9edc39c6260705"} Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.318125 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.410348 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-796zb\" (UniqueName: \"kubernetes.io/projected/5e5b7b28-ee18-4b85-af9c-030cab7da493-kube-api-access-796zb\") pod \"5e5b7b28-ee18-4b85-af9c-030cab7da493\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.410571 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-extra-config-0\") pod \"5e5b7b28-ee18-4b85-af9c-030cab7da493\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.410604 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-migration-ssh-key-1\") pod \"5e5b7b28-ee18-4b85-af9c-030cab7da493\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.410630 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-cell1-compute-config-1\") pod \"5e5b7b28-ee18-4b85-af9c-030cab7da493\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.410690 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-cell1-compute-config-0\") pod \"5e5b7b28-ee18-4b85-af9c-030cab7da493\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.410737 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-inventory\") pod \"5e5b7b28-ee18-4b85-af9c-030cab7da493\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.410765 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-migration-ssh-key-0\") pod \"5e5b7b28-ee18-4b85-af9c-030cab7da493\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.410794 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-combined-ca-bundle\") pod \"5e5b7b28-ee18-4b85-af9c-030cab7da493\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.410845 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-ssh-key\") pod \"5e5b7b28-ee18-4b85-af9c-030cab7da493\" (UID: \"5e5b7b28-ee18-4b85-af9c-030cab7da493\") " Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.436550 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "5e5b7b28-ee18-4b85-af9c-030cab7da493" (UID: "5e5b7b28-ee18-4b85-af9c-030cab7da493"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.436586 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e5b7b28-ee18-4b85-af9c-030cab7da493-kube-api-access-796zb" (OuterVolumeSpecName: "kube-api-access-796zb") pod "5e5b7b28-ee18-4b85-af9c-030cab7da493" (UID: "5e5b7b28-ee18-4b85-af9c-030cab7da493"). InnerVolumeSpecName "kube-api-access-796zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.452325 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "5e5b7b28-ee18-4b85-af9c-030cab7da493" (UID: "5e5b7b28-ee18-4b85-af9c-030cab7da493"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.456137 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "5e5b7b28-ee18-4b85-af9c-030cab7da493" (UID: "5e5b7b28-ee18-4b85-af9c-030cab7da493"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.456836 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "5e5b7b28-ee18-4b85-af9c-030cab7da493" (UID: "5e5b7b28-ee18-4b85-af9c-030cab7da493"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.458286 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5e5b7b28-ee18-4b85-af9c-030cab7da493" (UID: "5e5b7b28-ee18-4b85-af9c-030cab7da493"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.466679 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-inventory" (OuterVolumeSpecName: "inventory") pod "5e5b7b28-ee18-4b85-af9c-030cab7da493" (UID: "5e5b7b28-ee18-4b85-af9c-030cab7da493"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.471480 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "5e5b7b28-ee18-4b85-af9c-030cab7da493" (UID: "5e5b7b28-ee18-4b85-af9c-030cab7da493"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.492382 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "5e5b7b28-ee18-4b85-af9c-030cab7da493" (UID: "5e5b7b28-ee18-4b85-af9c-030cab7da493"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.516347 4685 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.516382 4685 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.516392 4685 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.516403 4685 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.516412 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.516421 4685 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.516429 4685 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.516438 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e5b7b28-ee18-4b85-af9c-030cab7da493-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.516447 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-796zb\" (UniqueName: \"kubernetes.io/projected/5e5b7b28-ee18-4b85-af9c-030cab7da493-kube-api-access-796zb\") on node \"crc\" DevicePath \"\"" Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.890895 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" event={"ID":"5e5b7b28-ee18-4b85-af9c-030cab7da493","Type":"ContainerDied","Data":"606a66da1cee419df6eb14979eadb59edf475f42a22b87d2fdb37a6b6f8f7d85"} Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.890974 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="606a66da1cee419df6eb14979eadb59edf475f42a22b87d2fdb37a6b6f8f7d85" Oct 13 09:28:08 crc kubenswrapper[4685]: I1013 09:28:08.891804 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-f2kmw" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.048394 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg"] Oct 13 09:28:09 crc kubenswrapper[4685]: E1013 09:28:09.049492 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5b7b28-ee18-4b85-af9c-030cab7da493" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.049596 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5b7b28-ee18-4b85-af9c-030cab7da493" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.049906 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e5b7b28-ee18-4b85-af9c-030cab7da493" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.050956 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.053014 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.054366 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.054508 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.057343 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.059120 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hcn4z" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.067159 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg"] Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.133823 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.134112 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.134238 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.134400 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.134451 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.134545 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.134571 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhswg\" (UniqueName: \"kubernetes.io/projected/e955f1da-bde6-47d5-8cdf-fe0d7182c081-kube-api-access-jhswg\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.237137 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.237202 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhswg\" (UniqueName: \"kubernetes.io/projected/e955f1da-bde6-47d5-8cdf-fe0d7182c081-kube-api-access-jhswg\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.237300 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.237369 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.237440 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.237484 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.237518 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.243471 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.243513 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.245348 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.250480 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.250988 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.255690 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.256557 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhswg\" (UniqueName: \"kubernetes.io/projected/e955f1da-bde6-47d5-8cdf-fe0d7182c081-kube-api-access-jhswg\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.418987 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:28:09 crc kubenswrapper[4685]: I1013 09:28:09.993604 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg"] Oct 13 09:28:10 crc kubenswrapper[4685]: I1013 09:28:10.005327 4685 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 13 09:28:10 crc kubenswrapper[4685]: I1013 09:28:10.915038 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" event={"ID":"e955f1da-bde6-47d5-8cdf-fe0d7182c081","Type":"ContainerStarted","Data":"2b0b224a3cf819d933a4be7286887bbf4e1047de1093960c6d487e658181e619"} Oct 13 09:28:10 crc kubenswrapper[4685]: I1013 09:28:10.915472 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" event={"ID":"e955f1da-bde6-47d5-8cdf-fe0d7182c081","Type":"ContainerStarted","Data":"eedc752cdfe985ee81cb7c64fb434a55221020a643614cb1a2a315a10af2f345"} Oct 13 09:28:10 crc kubenswrapper[4685]: I1013 09:28:10.940489 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" podStartSLOduration=1.433216001 podStartE2EDuration="1.940464143s" podCreationTimestamp="2025-10-13 09:28:09 +0000 UTC" firstStartedPulling="2025-10-13 09:28:10.005145336 +0000 UTC m=+2615.153021097" lastFinishedPulling="2025-10-13 09:28:10.512393478 +0000 UTC m=+2615.660269239" observedRunningTime="2025-10-13 09:28:10.931140222 +0000 UTC m=+2616.079015983" watchObservedRunningTime="2025-10-13 09:28:10.940464143 +0000 UTC m=+2616.088339904" Oct 13 09:28:13 crc kubenswrapper[4685]: I1013 09:28:13.502814 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:28:13 crc kubenswrapper[4685]: E1013 09:28:13.504752 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:28:27 crc kubenswrapper[4685]: I1013 09:28:27.503229 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:28:27 crc kubenswrapper[4685]: E1013 09:28:27.503962 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:28:42 crc kubenswrapper[4685]: I1013 09:28:42.509624 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:28:42 crc kubenswrapper[4685]: E1013 09:28:42.510626 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:28:53 crc kubenswrapper[4685]: I1013 09:28:53.502602 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:28:53 crc kubenswrapper[4685]: E1013 09:28:53.503356 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:29:05 crc kubenswrapper[4685]: I1013 09:29:05.513072 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:29:05 crc kubenswrapper[4685]: E1013 09:29:05.515280 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:29:14 crc kubenswrapper[4685]: I1013 09:29:14.937267 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v4m9t"] Oct 13 09:29:14 crc kubenswrapper[4685]: I1013 09:29:14.940419 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4m9t" Oct 13 09:29:14 crc kubenswrapper[4685]: I1013 09:29:14.951660 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v4m9t"] Oct 13 09:29:15 crc kubenswrapper[4685]: I1013 09:29:14.999806 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nnjj\" (UniqueName: \"kubernetes.io/projected/95b5dce1-9d4f-42d2-bda4-2df6de83c3f7-kube-api-access-6nnjj\") pod \"certified-operators-v4m9t\" (UID: \"95b5dce1-9d4f-42d2-bda4-2df6de83c3f7\") " pod="openshift-marketplace/certified-operators-v4m9t" Oct 13 09:29:15 crc kubenswrapper[4685]: I1013 09:29:14.999998 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95b5dce1-9d4f-42d2-bda4-2df6de83c3f7-catalog-content\") pod \"certified-operators-v4m9t\" (UID: \"95b5dce1-9d4f-42d2-bda4-2df6de83c3f7\") " pod="openshift-marketplace/certified-operators-v4m9t" Oct 13 09:29:15 crc kubenswrapper[4685]: I1013 09:29:15.000199 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95b5dce1-9d4f-42d2-bda4-2df6de83c3f7-utilities\") pod \"certified-operators-v4m9t\" (UID: \"95b5dce1-9d4f-42d2-bda4-2df6de83c3f7\") " pod="openshift-marketplace/certified-operators-v4m9t" Oct 13 09:29:15 crc kubenswrapper[4685]: I1013 09:29:15.101850 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95b5dce1-9d4f-42d2-bda4-2df6de83c3f7-catalog-content\") pod \"certified-operators-v4m9t\" (UID: \"95b5dce1-9d4f-42d2-bda4-2df6de83c3f7\") " pod="openshift-marketplace/certified-operators-v4m9t" Oct 13 09:29:15 crc kubenswrapper[4685]: I1013 09:29:15.102180 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95b5dce1-9d4f-42d2-bda4-2df6de83c3f7-utilities\") pod \"certified-operators-v4m9t\" (UID: \"95b5dce1-9d4f-42d2-bda4-2df6de83c3f7\") " pod="openshift-marketplace/certified-operators-v4m9t" Oct 13 09:29:15 crc kubenswrapper[4685]: I1013 09:29:15.102304 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nnjj\" (UniqueName: \"kubernetes.io/projected/95b5dce1-9d4f-42d2-bda4-2df6de83c3f7-kube-api-access-6nnjj\") pod \"certified-operators-v4m9t\" (UID: \"95b5dce1-9d4f-42d2-bda4-2df6de83c3f7\") " pod="openshift-marketplace/certified-operators-v4m9t" Oct 13 09:29:15 crc kubenswrapper[4685]: I1013 09:29:15.102406 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95b5dce1-9d4f-42d2-bda4-2df6de83c3f7-catalog-content\") pod \"certified-operators-v4m9t\" (UID: \"95b5dce1-9d4f-42d2-bda4-2df6de83c3f7\") " pod="openshift-marketplace/certified-operators-v4m9t" Oct 13 09:29:15 crc kubenswrapper[4685]: I1013 09:29:15.102424 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95b5dce1-9d4f-42d2-bda4-2df6de83c3f7-utilities\") pod \"certified-operators-v4m9t\" (UID: \"95b5dce1-9d4f-42d2-bda4-2df6de83c3f7\") " pod="openshift-marketplace/certified-operators-v4m9t" Oct 13 09:29:15 crc kubenswrapper[4685]: I1013 09:29:15.128155 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nnjj\" (UniqueName: \"kubernetes.io/projected/95b5dce1-9d4f-42d2-bda4-2df6de83c3f7-kube-api-access-6nnjj\") pod \"certified-operators-v4m9t\" (UID: \"95b5dce1-9d4f-42d2-bda4-2df6de83c3f7\") " pod="openshift-marketplace/certified-operators-v4m9t" Oct 13 09:29:15 crc kubenswrapper[4685]: I1013 09:29:15.285904 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4m9t" Oct 13 09:29:15 crc kubenswrapper[4685]: I1013 09:29:15.951573 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v4m9t"] Oct 13 09:29:16 crc kubenswrapper[4685]: I1013 09:29:16.552081 4685 generic.go:334] "Generic (PLEG): container finished" podID="95b5dce1-9d4f-42d2-bda4-2df6de83c3f7" containerID="f8d1ba93a9041c3f91501e251d4adc581db8579547449f975bd48b37d653d041" exitCode=0 Oct 13 09:29:16 crc kubenswrapper[4685]: I1013 09:29:16.552175 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4m9t" event={"ID":"95b5dce1-9d4f-42d2-bda4-2df6de83c3f7","Type":"ContainerDied","Data":"f8d1ba93a9041c3f91501e251d4adc581db8579547449f975bd48b37d653d041"} Oct 13 09:29:16 crc kubenswrapper[4685]: I1013 09:29:16.553066 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4m9t" event={"ID":"95b5dce1-9d4f-42d2-bda4-2df6de83c3f7","Type":"ContainerStarted","Data":"055c04260274f8511250a6bdb62b3fadd7ddcf2f47f101aba94205770f71f877"} Oct 13 09:29:17 crc kubenswrapper[4685]: I1013 09:29:17.563731 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4m9t" event={"ID":"95b5dce1-9d4f-42d2-bda4-2df6de83c3f7","Type":"ContainerStarted","Data":"aa5c72a3286e24f2063a36d43ca71501128718dc64eb46ef36745933c3005917"} Oct 13 09:29:20 crc kubenswrapper[4685]: I1013 09:29:20.503415 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:29:20 crc kubenswrapper[4685]: E1013 09:29:20.504231 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:29:20 crc kubenswrapper[4685]: I1013 09:29:20.589754 4685 generic.go:334] "Generic (PLEG): container finished" podID="95b5dce1-9d4f-42d2-bda4-2df6de83c3f7" containerID="aa5c72a3286e24f2063a36d43ca71501128718dc64eb46ef36745933c3005917" exitCode=0 Oct 13 09:29:20 crc kubenswrapper[4685]: I1013 09:29:20.589999 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4m9t" event={"ID":"95b5dce1-9d4f-42d2-bda4-2df6de83c3f7","Type":"ContainerDied","Data":"aa5c72a3286e24f2063a36d43ca71501128718dc64eb46ef36745933c3005917"} Oct 13 09:29:22 crc kubenswrapper[4685]: I1013 09:29:22.616218 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4m9t" event={"ID":"95b5dce1-9d4f-42d2-bda4-2df6de83c3f7","Type":"ContainerStarted","Data":"2c0035647e3ab16da7f812ceef9ff127357cacd6f7412c92f893fc1dc596adcd"} Oct 13 09:29:22 crc kubenswrapper[4685]: I1013 09:29:22.647689 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v4m9t" podStartSLOduration=3.641009586 podStartE2EDuration="8.647666942s" podCreationTimestamp="2025-10-13 09:29:14 +0000 UTC" firstStartedPulling="2025-10-13 09:29:16.554207219 +0000 UTC m=+2681.702082980" lastFinishedPulling="2025-10-13 09:29:21.560864575 +0000 UTC m=+2686.708740336" observedRunningTime="2025-10-13 09:29:22.638361581 +0000 UTC m=+2687.786237362" watchObservedRunningTime="2025-10-13 09:29:22.647666942 +0000 UTC m=+2687.795542703" Oct 13 09:29:25 crc kubenswrapper[4685]: I1013 09:29:25.286455 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v4m9t" Oct 13 09:29:25 crc kubenswrapper[4685]: I1013 09:29:25.288132 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v4m9t" Oct 13 09:29:25 crc kubenswrapper[4685]: I1013 09:29:25.366798 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v4m9t" Oct 13 09:29:26 crc kubenswrapper[4685]: I1013 09:29:26.703111 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v4m9t" Oct 13 09:29:26 crc kubenswrapper[4685]: I1013 09:29:26.758794 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v4m9t"] Oct 13 09:29:28 crc kubenswrapper[4685]: I1013 09:29:28.671613 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v4m9t" podUID="95b5dce1-9d4f-42d2-bda4-2df6de83c3f7" containerName="registry-server" containerID="cri-o://2c0035647e3ab16da7f812ceef9ff127357cacd6f7412c92f893fc1dc596adcd" gracePeriod=2 Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.147298 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4m9t" Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.317105 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95b5dce1-9d4f-42d2-bda4-2df6de83c3f7-utilities\") pod \"95b5dce1-9d4f-42d2-bda4-2df6de83c3f7\" (UID: \"95b5dce1-9d4f-42d2-bda4-2df6de83c3f7\") " Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.317266 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95b5dce1-9d4f-42d2-bda4-2df6de83c3f7-catalog-content\") pod \"95b5dce1-9d4f-42d2-bda4-2df6de83c3f7\" (UID: \"95b5dce1-9d4f-42d2-bda4-2df6de83c3f7\") " Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.317381 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nnjj\" (UniqueName: \"kubernetes.io/projected/95b5dce1-9d4f-42d2-bda4-2df6de83c3f7-kube-api-access-6nnjj\") pod \"95b5dce1-9d4f-42d2-bda4-2df6de83c3f7\" (UID: \"95b5dce1-9d4f-42d2-bda4-2df6de83c3f7\") " Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.318527 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95b5dce1-9d4f-42d2-bda4-2df6de83c3f7-utilities" (OuterVolumeSpecName: "utilities") pod "95b5dce1-9d4f-42d2-bda4-2df6de83c3f7" (UID: "95b5dce1-9d4f-42d2-bda4-2df6de83c3f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.342213 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95b5dce1-9d4f-42d2-bda4-2df6de83c3f7-kube-api-access-6nnjj" (OuterVolumeSpecName: "kube-api-access-6nnjj") pod "95b5dce1-9d4f-42d2-bda4-2df6de83c3f7" (UID: "95b5dce1-9d4f-42d2-bda4-2df6de83c3f7"). InnerVolumeSpecName "kube-api-access-6nnjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.377226 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95b5dce1-9d4f-42d2-bda4-2df6de83c3f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95b5dce1-9d4f-42d2-bda4-2df6de83c3f7" (UID: "95b5dce1-9d4f-42d2-bda4-2df6de83c3f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.420248 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95b5dce1-9d4f-42d2-bda4-2df6de83c3f7-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.420321 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95b5dce1-9d4f-42d2-bda4-2df6de83c3f7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.420353 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nnjj\" (UniqueName: \"kubernetes.io/projected/95b5dce1-9d4f-42d2-bda4-2df6de83c3f7-kube-api-access-6nnjj\") on node \"crc\" DevicePath \"\"" Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.687716 4685 generic.go:334] "Generic (PLEG): container finished" podID="95b5dce1-9d4f-42d2-bda4-2df6de83c3f7" containerID="2c0035647e3ab16da7f812ceef9ff127357cacd6f7412c92f893fc1dc596adcd" exitCode=0 Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.687767 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4m9t" event={"ID":"95b5dce1-9d4f-42d2-bda4-2df6de83c3f7","Type":"ContainerDied","Data":"2c0035647e3ab16da7f812ceef9ff127357cacd6f7412c92f893fc1dc596adcd"} Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.687824 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4m9t" Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.687843 4685 scope.go:117] "RemoveContainer" containerID="2c0035647e3ab16da7f812ceef9ff127357cacd6f7412c92f893fc1dc596adcd" Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.687830 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4m9t" event={"ID":"95b5dce1-9d4f-42d2-bda4-2df6de83c3f7","Type":"ContainerDied","Data":"055c04260274f8511250a6bdb62b3fadd7ddcf2f47f101aba94205770f71f877"} Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.726000 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v4m9t"] Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.740316 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v4m9t"] Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.744527 4685 scope.go:117] "RemoveContainer" containerID="aa5c72a3286e24f2063a36d43ca71501128718dc64eb46ef36745933c3005917" Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.771431 4685 scope.go:117] "RemoveContainer" containerID="f8d1ba93a9041c3f91501e251d4adc581db8579547449f975bd48b37d653d041" Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.810630 4685 scope.go:117] "RemoveContainer" containerID="2c0035647e3ab16da7f812ceef9ff127357cacd6f7412c92f893fc1dc596adcd" Oct 13 09:29:29 crc kubenswrapper[4685]: E1013 09:29:29.811169 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c0035647e3ab16da7f812ceef9ff127357cacd6f7412c92f893fc1dc596adcd\": container with ID starting with 2c0035647e3ab16da7f812ceef9ff127357cacd6f7412c92f893fc1dc596adcd not found: ID does not exist" containerID="2c0035647e3ab16da7f812ceef9ff127357cacd6f7412c92f893fc1dc596adcd" Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.811220 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c0035647e3ab16da7f812ceef9ff127357cacd6f7412c92f893fc1dc596adcd"} err="failed to get container status \"2c0035647e3ab16da7f812ceef9ff127357cacd6f7412c92f893fc1dc596adcd\": rpc error: code = NotFound desc = could not find container \"2c0035647e3ab16da7f812ceef9ff127357cacd6f7412c92f893fc1dc596adcd\": container with ID starting with 2c0035647e3ab16da7f812ceef9ff127357cacd6f7412c92f893fc1dc596adcd not found: ID does not exist" Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.811251 4685 scope.go:117] "RemoveContainer" containerID="aa5c72a3286e24f2063a36d43ca71501128718dc64eb46ef36745933c3005917" Oct 13 09:29:29 crc kubenswrapper[4685]: E1013 09:29:29.811635 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa5c72a3286e24f2063a36d43ca71501128718dc64eb46ef36745933c3005917\": container with ID starting with aa5c72a3286e24f2063a36d43ca71501128718dc64eb46ef36745933c3005917 not found: ID does not exist" containerID="aa5c72a3286e24f2063a36d43ca71501128718dc64eb46ef36745933c3005917" Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.811667 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa5c72a3286e24f2063a36d43ca71501128718dc64eb46ef36745933c3005917"} err="failed to get container status \"aa5c72a3286e24f2063a36d43ca71501128718dc64eb46ef36745933c3005917\": rpc error: code = NotFound desc = could not find container \"aa5c72a3286e24f2063a36d43ca71501128718dc64eb46ef36745933c3005917\": container with ID starting with aa5c72a3286e24f2063a36d43ca71501128718dc64eb46ef36745933c3005917 not found: ID does not exist" Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.811688 4685 scope.go:117] "RemoveContainer" containerID="f8d1ba93a9041c3f91501e251d4adc581db8579547449f975bd48b37d653d041" Oct 13 09:29:29 crc kubenswrapper[4685]: E1013 09:29:29.811934 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8d1ba93a9041c3f91501e251d4adc581db8579547449f975bd48b37d653d041\": container with ID starting with f8d1ba93a9041c3f91501e251d4adc581db8579547449f975bd48b37d653d041 not found: ID does not exist" containerID="f8d1ba93a9041c3f91501e251d4adc581db8579547449f975bd48b37d653d041" Oct 13 09:29:29 crc kubenswrapper[4685]: I1013 09:29:29.811961 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8d1ba93a9041c3f91501e251d4adc581db8579547449f975bd48b37d653d041"} err="failed to get container status \"f8d1ba93a9041c3f91501e251d4adc581db8579547449f975bd48b37d653d041\": rpc error: code = NotFound desc = could not find container \"f8d1ba93a9041c3f91501e251d4adc581db8579547449f975bd48b37d653d041\": container with ID starting with f8d1ba93a9041c3f91501e251d4adc581db8579547449f975bd48b37d653d041 not found: ID does not exist" Oct 13 09:29:31 crc kubenswrapper[4685]: I1013 09:29:31.517618 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95b5dce1-9d4f-42d2-bda4-2df6de83c3f7" path="/var/lib/kubelet/pods/95b5dce1-9d4f-42d2-bda4-2df6de83c3f7/volumes" Oct 13 09:29:34 crc kubenswrapper[4685]: I1013 09:29:34.503688 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:29:34 crc kubenswrapper[4685]: E1013 09:29:34.504524 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:29:48 crc kubenswrapper[4685]: I1013 09:29:48.502717 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:29:48 crc kubenswrapper[4685]: E1013 09:29:48.503483 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:30:00 crc kubenswrapper[4685]: I1013 09:30:00.146794 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339130-lmt4z"] Oct 13 09:30:00 crc kubenswrapper[4685]: E1013 09:30:00.147953 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b5dce1-9d4f-42d2-bda4-2df6de83c3f7" containerName="extract-utilities" Oct 13 09:30:00 crc kubenswrapper[4685]: I1013 09:30:00.147973 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b5dce1-9d4f-42d2-bda4-2df6de83c3f7" containerName="extract-utilities" Oct 13 09:30:00 crc kubenswrapper[4685]: E1013 09:30:00.148008 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b5dce1-9d4f-42d2-bda4-2df6de83c3f7" containerName="extract-content" Oct 13 09:30:00 crc kubenswrapper[4685]: I1013 09:30:00.148016 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b5dce1-9d4f-42d2-bda4-2df6de83c3f7" containerName="extract-content" Oct 13 09:30:00 crc kubenswrapper[4685]: E1013 09:30:00.148050 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b5dce1-9d4f-42d2-bda4-2df6de83c3f7" containerName="registry-server" Oct 13 09:30:00 crc kubenswrapper[4685]: I1013 09:30:00.148059 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b5dce1-9d4f-42d2-bda4-2df6de83c3f7" containerName="registry-server" Oct 13 09:30:00 crc kubenswrapper[4685]: I1013 09:30:00.148309 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b5dce1-9d4f-42d2-bda4-2df6de83c3f7" containerName="registry-server" Oct 13 09:30:00 crc kubenswrapper[4685]: I1013 09:30:00.149098 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339130-lmt4z" Oct 13 09:30:00 crc kubenswrapper[4685]: I1013 09:30:00.155542 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 13 09:30:00 crc kubenswrapper[4685]: I1013 09:30:00.155555 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 13 09:30:00 crc kubenswrapper[4685]: I1013 09:30:00.158072 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339130-lmt4z"] Oct 13 09:30:00 crc kubenswrapper[4685]: I1013 09:30:00.346734 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b86048e2-c562-406b-a033-6846a1a9b9f4-secret-volume\") pod \"collect-profiles-29339130-lmt4z\" (UID: \"b86048e2-c562-406b-a033-6846a1a9b9f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339130-lmt4z" Oct 13 09:30:00 crc kubenswrapper[4685]: I1013 09:30:00.346775 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b86048e2-c562-406b-a033-6846a1a9b9f4-config-volume\") pod \"collect-profiles-29339130-lmt4z\" (UID: \"b86048e2-c562-406b-a033-6846a1a9b9f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339130-lmt4z" Oct 13 09:30:00 crc kubenswrapper[4685]: I1013 09:30:00.346814 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcjjt\" (UniqueName: \"kubernetes.io/projected/b86048e2-c562-406b-a033-6846a1a9b9f4-kube-api-access-mcjjt\") pod \"collect-profiles-29339130-lmt4z\" (UID: \"b86048e2-c562-406b-a033-6846a1a9b9f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339130-lmt4z" Oct 13 09:30:00 crc kubenswrapper[4685]: I1013 09:30:00.449137 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b86048e2-c562-406b-a033-6846a1a9b9f4-secret-volume\") pod \"collect-profiles-29339130-lmt4z\" (UID: \"b86048e2-c562-406b-a033-6846a1a9b9f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339130-lmt4z" Oct 13 09:30:00 crc kubenswrapper[4685]: I1013 09:30:00.449214 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b86048e2-c562-406b-a033-6846a1a9b9f4-config-volume\") pod \"collect-profiles-29339130-lmt4z\" (UID: \"b86048e2-c562-406b-a033-6846a1a9b9f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339130-lmt4z" Oct 13 09:30:00 crc kubenswrapper[4685]: I1013 09:30:00.449282 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcjjt\" (UniqueName: \"kubernetes.io/projected/b86048e2-c562-406b-a033-6846a1a9b9f4-kube-api-access-mcjjt\") pod \"collect-profiles-29339130-lmt4z\" (UID: \"b86048e2-c562-406b-a033-6846a1a9b9f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339130-lmt4z" Oct 13 09:30:00 crc kubenswrapper[4685]: I1013 09:30:00.450377 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b86048e2-c562-406b-a033-6846a1a9b9f4-config-volume\") pod \"collect-profiles-29339130-lmt4z\" (UID: \"b86048e2-c562-406b-a033-6846a1a9b9f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339130-lmt4z" Oct 13 09:30:00 crc kubenswrapper[4685]: I1013 09:30:00.457968 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b86048e2-c562-406b-a033-6846a1a9b9f4-secret-volume\") pod \"collect-profiles-29339130-lmt4z\" (UID: \"b86048e2-c562-406b-a033-6846a1a9b9f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339130-lmt4z" Oct 13 09:30:00 crc kubenswrapper[4685]: I1013 09:30:00.466012 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcjjt\" (UniqueName: \"kubernetes.io/projected/b86048e2-c562-406b-a033-6846a1a9b9f4-kube-api-access-mcjjt\") pod \"collect-profiles-29339130-lmt4z\" (UID: \"b86048e2-c562-406b-a033-6846a1a9b9f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339130-lmt4z" Oct 13 09:30:00 crc kubenswrapper[4685]: I1013 09:30:00.469621 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339130-lmt4z" Oct 13 09:30:00 crc kubenswrapper[4685]: I1013 09:30:00.961645 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339130-lmt4z"] Oct 13 09:30:00 crc kubenswrapper[4685]: I1013 09:30:00.999124 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339130-lmt4z" event={"ID":"b86048e2-c562-406b-a033-6846a1a9b9f4","Type":"ContainerStarted","Data":"83d18c74cad3a2150a9c49e2bc217b74876574507cd6ff0d1b9c22a337441926"} Oct 13 09:30:02 crc kubenswrapper[4685]: I1013 09:30:02.008982 4685 generic.go:334] "Generic (PLEG): container finished" podID="b86048e2-c562-406b-a033-6846a1a9b9f4" containerID="78a49904d970109bc9c0eddb4ff1b429db693754450b2f9a2e91e0f07e8dd446" exitCode=0 Oct 13 09:30:02 crc kubenswrapper[4685]: I1013 09:30:02.009484 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339130-lmt4z" event={"ID":"b86048e2-c562-406b-a033-6846a1a9b9f4","Type":"ContainerDied","Data":"78a49904d970109bc9c0eddb4ff1b429db693754450b2f9a2e91e0f07e8dd446"} Oct 13 09:30:02 crc kubenswrapper[4685]: I1013 09:30:02.502443 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:30:03 crc kubenswrapper[4685]: I1013 09:30:03.025576 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerStarted","Data":"f1a9443768e1ae7075cbb2eb7a5ba6388414738c043869b315940a2a75bdea21"} Oct 13 09:30:03 crc kubenswrapper[4685]: I1013 09:30:03.387646 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339130-lmt4z" Oct 13 09:30:03 crc kubenswrapper[4685]: I1013 09:30:03.509854 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcjjt\" (UniqueName: \"kubernetes.io/projected/b86048e2-c562-406b-a033-6846a1a9b9f4-kube-api-access-mcjjt\") pod \"b86048e2-c562-406b-a033-6846a1a9b9f4\" (UID: \"b86048e2-c562-406b-a033-6846a1a9b9f4\") " Oct 13 09:30:03 crc kubenswrapper[4685]: I1013 09:30:03.509890 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b86048e2-c562-406b-a033-6846a1a9b9f4-secret-volume\") pod \"b86048e2-c562-406b-a033-6846a1a9b9f4\" (UID: \"b86048e2-c562-406b-a033-6846a1a9b9f4\") " Oct 13 09:30:03 crc kubenswrapper[4685]: I1013 09:30:03.510018 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b86048e2-c562-406b-a033-6846a1a9b9f4-config-volume\") pod \"b86048e2-c562-406b-a033-6846a1a9b9f4\" (UID: \"b86048e2-c562-406b-a033-6846a1a9b9f4\") " Oct 13 09:30:03 crc kubenswrapper[4685]: I1013 09:30:03.511138 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b86048e2-c562-406b-a033-6846a1a9b9f4-config-volume" (OuterVolumeSpecName: "config-volume") pod "b86048e2-c562-406b-a033-6846a1a9b9f4" (UID: "b86048e2-c562-406b-a033-6846a1a9b9f4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:30:03 crc kubenswrapper[4685]: I1013 09:30:03.517839 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b86048e2-c562-406b-a033-6846a1a9b9f4-kube-api-access-mcjjt" (OuterVolumeSpecName: "kube-api-access-mcjjt") pod "b86048e2-c562-406b-a033-6846a1a9b9f4" (UID: "b86048e2-c562-406b-a033-6846a1a9b9f4"). InnerVolumeSpecName "kube-api-access-mcjjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:30:03 crc kubenswrapper[4685]: I1013 09:30:03.523627 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b86048e2-c562-406b-a033-6846a1a9b9f4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b86048e2-c562-406b-a033-6846a1a9b9f4" (UID: "b86048e2-c562-406b-a033-6846a1a9b9f4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:30:03 crc kubenswrapper[4685]: I1013 09:30:03.612250 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcjjt\" (UniqueName: \"kubernetes.io/projected/b86048e2-c562-406b-a033-6846a1a9b9f4-kube-api-access-mcjjt\") on node \"crc\" DevicePath \"\"" Oct 13 09:30:03 crc kubenswrapper[4685]: I1013 09:30:03.612296 4685 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b86048e2-c562-406b-a033-6846a1a9b9f4-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 13 09:30:03 crc kubenswrapper[4685]: I1013 09:30:03.612309 4685 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b86048e2-c562-406b-a033-6846a1a9b9f4-config-volume\") on node \"crc\" DevicePath \"\"" Oct 13 09:30:04 crc kubenswrapper[4685]: I1013 09:30:04.037698 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339130-lmt4z" event={"ID":"b86048e2-c562-406b-a033-6846a1a9b9f4","Type":"ContainerDied","Data":"83d18c74cad3a2150a9c49e2bc217b74876574507cd6ff0d1b9c22a337441926"} Oct 13 09:30:04 crc kubenswrapper[4685]: I1013 09:30:04.038001 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83d18c74cad3a2150a9c49e2bc217b74876574507cd6ff0d1b9c22a337441926" Oct 13 09:30:04 crc kubenswrapper[4685]: I1013 09:30:04.037764 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339130-lmt4z" Oct 13 09:30:04 crc kubenswrapper[4685]: I1013 09:30:04.468199 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn"] Oct 13 09:30:04 crc kubenswrapper[4685]: I1013 09:30:04.475496 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339085-84tsn"] Oct 13 09:30:05 crc kubenswrapper[4685]: I1013 09:30:05.521021 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3672a087-9d9b-45ee-8f80-686959b395d6" path="/var/lib/kubelet/pods/3672a087-9d9b-45ee-8f80-686959b395d6/volumes" Oct 13 09:30:12 crc kubenswrapper[4685]: I1013 09:30:12.105579 4685 generic.go:334] "Generic (PLEG): container finished" podID="b001d17a-1aea-44ba-86c5-ba6b312156c1" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" exitCode=1 Oct 13 09:30:12 crc kubenswrapper[4685]: I1013 09:30:12.105757 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" event={"ID":"b001d17a-1aea-44ba-86c5-ba6b312156c1","Type":"ContainerDied","Data":"69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b"} Oct 13 09:30:12 crc kubenswrapper[4685]: I1013 09:30:12.106202 4685 scope.go:117] "RemoveContainer" containerID="6fe2a2d2c5bb6ca13a66f41b5c31e8c9383694170a780b3627dfd3d787a8306e" Oct 13 09:30:12 crc kubenswrapper[4685]: I1013 09:30:12.106977 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:30:12 crc kubenswrapper[4685]: E1013 09:30:12.107403 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:30:13 crc kubenswrapper[4685]: I1013 09:30:13.442957 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:30:13 crc kubenswrapper[4685]: I1013 09:30:13.444690 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:30:13 crc kubenswrapper[4685]: E1013 09:30:13.445011 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:30:23 crc kubenswrapper[4685]: I1013 09:30:23.442545 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:30:23 crc kubenswrapper[4685]: I1013 09:30:23.444003 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:30:23 crc kubenswrapper[4685]: E1013 09:30:23.444278 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:30:31 crc kubenswrapper[4685]: I1013 09:30:31.269670 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xr297"] Oct 13 09:30:31 crc kubenswrapper[4685]: E1013 09:30:31.271211 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b86048e2-c562-406b-a033-6846a1a9b9f4" containerName="collect-profiles" Oct 13 09:30:31 crc kubenswrapper[4685]: I1013 09:30:31.271230 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="b86048e2-c562-406b-a033-6846a1a9b9f4" containerName="collect-profiles" Oct 13 09:30:31 crc kubenswrapper[4685]: I1013 09:30:31.271497 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="b86048e2-c562-406b-a033-6846a1a9b9f4" containerName="collect-profiles" Oct 13 09:30:31 crc kubenswrapper[4685]: I1013 09:30:31.273215 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xr297" Oct 13 09:30:31 crc kubenswrapper[4685]: I1013 09:30:31.293335 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xr297"] Oct 13 09:30:31 crc kubenswrapper[4685]: I1013 09:30:31.435984 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dd114e4-8912-4bbd-831c-40d9621154a4-utilities\") pod \"redhat-operators-xr297\" (UID: \"0dd114e4-8912-4bbd-831c-40d9621154a4\") " pod="openshift-marketplace/redhat-operators-xr297" Oct 13 09:30:31 crc kubenswrapper[4685]: I1013 09:30:31.436054 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dd114e4-8912-4bbd-831c-40d9621154a4-catalog-content\") pod \"redhat-operators-xr297\" (UID: \"0dd114e4-8912-4bbd-831c-40d9621154a4\") " pod="openshift-marketplace/redhat-operators-xr297" Oct 13 09:30:31 crc kubenswrapper[4685]: I1013 09:30:31.436233 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms4sm\" (UniqueName: \"kubernetes.io/projected/0dd114e4-8912-4bbd-831c-40d9621154a4-kube-api-access-ms4sm\") pod \"redhat-operators-xr297\" (UID: \"0dd114e4-8912-4bbd-831c-40d9621154a4\") " pod="openshift-marketplace/redhat-operators-xr297" Oct 13 09:30:31 crc kubenswrapper[4685]: I1013 09:30:31.539123 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms4sm\" (UniqueName: \"kubernetes.io/projected/0dd114e4-8912-4bbd-831c-40d9621154a4-kube-api-access-ms4sm\") pod \"redhat-operators-xr297\" (UID: \"0dd114e4-8912-4bbd-831c-40d9621154a4\") " pod="openshift-marketplace/redhat-operators-xr297" Oct 13 09:30:31 crc kubenswrapper[4685]: I1013 09:30:31.539605 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dd114e4-8912-4bbd-831c-40d9621154a4-utilities\") pod \"redhat-operators-xr297\" (UID: \"0dd114e4-8912-4bbd-831c-40d9621154a4\") " pod="openshift-marketplace/redhat-operators-xr297" Oct 13 09:30:31 crc kubenswrapper[4685]: I1013 09:30:31.539763 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dd114e4-8912-4bbd-831c-40d9621154a4-catalog-content\") pod \"redhat-operators-xr297\" (UID: \"0dd114e4-8912-4bbd-831c-40d9621154a4\") " pod="openshift-marketplace/redhat-operators-xr297" Oct 13 09:30:31 crc kubenswrapper[4685]: I1013 09:30:31.540178 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dd114e4-8912-4bbd-831c-40d9621154a4-utilities\") pod \"redhat-operators-xr297\" (UID: \"0dd114e4-8912-4bbd-831c-40d9621154a4\") " pod="openshift-marketplace/redhat-operators-xr297" Oct 13 09:30:31 crc kubenswrapper[4685]: I1013 09:30:31.540239 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dd114e4-8912-4bbd-831c-40d9621154a4-catalog-content\") pod \"redhat-operators-xr297\" (UID: \"0dd114e4-8912-4bbd-831c-40d9621154a4\") " pod="openshift-marketplace/redhat-operators-xr297" Oct 13 09:30:31 crc kubenswrapper[4685]: I1013 09:30:31.560484 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms4sm\" (UniqueName: \"kubernetes.io/projected/0dd114e4-8912-4bbd-831c-40d9621154a4-kube-api-access-ms4sm\") pod \"redhat-operators-xr297\" (UID: \"0dd114e4-8912-4bbd-831c-40d9621154a4\") " pod="openshift-marketplace/redhat-operators-xr297" Oct 13 09:30:31 crc kubenswrapper[4685]: I1013 09:30:31.604022 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xr297" Oct 13 09:30:32 crc kubenswrapper[4685]: I1013 09:30:32.097700 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xr297"] Oct 13 09:30:32 crc kubenswrapper[4685]: I1013 09:30:32.303422 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xr297" event={"ID":"0dd114e4-8912-4bbd-831c-40d9621154a4","Type":"ContainerStarted","Data":"277d20e5d8b05111d50357698cd9c0a4a85c0cabd80cbcded068ecf8a0a9c1c5"} Oct 13 09:30:32 crc kubenswrapper[4685]: I1013 09:30:32.303471 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xr297" event={"ID":"0dd114e4-8912-4bbd-831c-40d9621154a4","Type":"ContainerStarted","Data":"5eac150afd1659ccde21b3d8017b5af51f1687367361c1e76ce13ef93e16711b"} Oct 13 09:30:33 crc kubenswrapper[4685]: I1013 09:30:33.312533 4685 generic.go:334] "Generic (PLEG): container finished" podID="0dd114e4-8912-4bbd-831c-40d9621154a4" containerID="277d20e5d8b05111d50357698cd9c0a4a85c0cabd80cbcded068ecf8a0a9c1c5" exitCode=0 Oct 13 09:30:33 crc kubenswrapper[4685]: I1013 09:30:33.312837 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xr297" event={"ID":"0dd114e4-8912-4bbd-831c-40d9621154a4","Type":"ContainerDied","Data":"277d20e5d8b05111d50357698cd9c0a4a85c0cabd80cbcded068ecf8a0a9c1c5"} Oct 13 09:30:34 crc kubenswrapper[4685]: I1013 09:30:34.503952 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:30:34 crc kubenswrapper[4685]: E1013 09:30:34.504851 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:30:35 crc kubenswrapper[4685]: I1013 09:30:35.334394 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xr297" event={"ID":"0dd114e4-8912-4bbd-831c-40d9621154a4","Type":"ContainerStarted","Data":"37900305fe9cd3eadb6aa7d95d08abf435250cba073f81fecd27fa177d6bc667"} Oct 13 09:30:38 crc kubenswrapper[4685]: I1013 09:30:38.363498 4685 generic.go:334] "Generic (PLEG): container finished" podID="0dd114e4-8912-4bbd-831c-40d9621154a4" containerID="37900305fe9cd3eadb6aa7d95d08abf435250cba073f81fecd27fa177d6bc667" exitCode=0 Oct 13 09:30:38 crc kubenswrapper[4685]: I1013 09:30:38.363583 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xr297" event={"ID":"0dd114e4-8912-4bbd-831c-40d9621154a4","Type":"ContainerDied","Data":"37900305fe9cd3eadb6aa7d95d08abf435250cba073f81fecd27fa177d6bc667"} Oct 13 09:30:39 crc kubenswrapper[4685]: I1013 09:30:39.377551 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xr297" event={"ID":"0dd114e4-8912-4bbd-831c-40d9621154a4","Type":"ContainerStarted","Data":"085b545711ae72e9221813beb3387fa5a9548bf87ae4d6ebf130c1145b652fae"} Oct 13 09:30:41 crc kubenswrapper[4685]: I1013 09:30:41.606560 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xr297" Oct 13 09:30:41 crc kubenswrapper[4685]: I1013 09:30:41.607003 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xr297" Oct 13 09:30:42 crc kubenswrapper[4685]: I1013 09:30:42.680748 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xr297" podUID="0dd114e4-8912-4bbd-831c-40d9621154a4" containerName="registry-server" probeResult="failure" output=< Oct 13 09:30:42 crc kubenswrapper[4685]: timeout: failed to connect service ":50051" within 1s Oct 13 09:30:42 crc kubenswrapper[4685]: > Oct 13 09:30:47 crc kubenswrapper[4685]: I1013 09:30:47.503035 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:30:47 crc kubenswrapper[4685]: E1013 09:30:47.503898 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:30:51 crc kubenswrapper[4685]: I1013 09:30:51.659641 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xr297" Oct 13 09:30:51 crc kubenswrapper[4685]: I1013 09:30:51.683293 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xr297" podStartSLOduration=15.170044263 podStartE2EDuration="20.683272965s" podCreationTimestamp="2025-10-13 09:30:31 +0000 UTC" firstStartedPulling="2025-10-13 09:30:33.31493502 +0000 UTC m=+2758.462810791" lastFinishedPulling="2025-10-13 09:30:38.828163732 +0000 UTC m=+2763.976039493" observedRunningTime="2025-10-13 09:30:39.401525508 +0000 UTC m=+2764.549401269" watchObservedRunningTime="2025-10-13 09:30:51.683272965 +0000 UTC m=+2776.831148726" Oct 13 09:30:51 crc kubenswrapper[4685]: I1013 09:30:51.723563 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xr297" Oct 13 09:30:51 crc kubenswrapper[4685]: I1013 09:30:51.903504 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xr297"] Oct 13 09:30:53 crc kubenswrapper[4685]: I1013 09:30:53.507965 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xr297" podUID="0dd114e4-8912-4bbd-831c-40d9621154a4" containerName="registry-server" containerID="cri-o://085b545711ae72e9221813beb3387fa5a9548bf87ae4d6ebf130c1145b652fae" gracePeriod=2 Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.120018 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xr297" Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.156541 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ms4sm\" (UniqueName: \"kubernetes.io/projected/0dd114e4-8912-4bbd-831c-40d9621154a4-kube-api-access-ms4sm\") pod \"0dd114e4-8912-4bbd-831c-40d9621154a4\" (UID: \"0dd114e4-8912-4bbd-831c-40d9621154a4\") " Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.156647 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dd114e4-8912-4bbd-831c-40d9621154a4-catalog-content\") pod \"0dd114e4-8912-4bbd-831c-40d9621154a4\" (UID: \"0dd114e4-8912-4bbd-831c-40d9621154a4\") " Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.156738 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dd114e4-8912-4bbd-831c-40d9621154a4-utilities\") pod \"0dd114e4-8912-4bbd-831c-40d9621154a4\" (UID: \"0dd114e4-8912-4bbd-831c-40d9621154a4\") " Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.158105 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dd114e4-8912-4bbd-831c-40d9621154a4-utilities" (OuterVolumeSpecName: "utilities") pod "0dd114e4-8912-4bbd-831c-40d9621154a4" (UID: "0dd114e4-8912-4bbd-831c-40d9621154a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.172295 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dd114e4-8912-4bbd-831c-40d9621154a4-kube-api-access-ms4sm" (OuterVolumeSpecName: "kube-api-access-ms4sm") pod "0dd114e4-8912-4bbd-831c-40d9621154a4" (UID: "0dd114e4-8912-4bbd-831c-40d9621154a4"). InnerVolumeSpecName "kube-api-access-ms4sm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.264102 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ms4sm\" (UniqueName: \"kubernetes.io/projected/0dd114e4-8912-4bbd-831c-40d9621154a4-kube-api-access-ms4sm\") on node \"crc\" DevicePath \"\"" Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.264137 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dd114e4-8912-4bbd-831c-40d9621154a4-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.284380 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dd114e4-8912-4bbd-831c-40d9621154a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0dd114e4-8912-4bbd-831c-40d9621154a4" (UID: "0dd114e4-8912-4bbd-831c-40d9621154a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.365576 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dd114e4-8912-4bbd-831c-40d9621154a4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.520175 4685 generic.go:334] "Generic (PLEG): container finished" podID="0dd114e4-8912-4bbd-831c-40d9621154a4" containerID="085b545711ae72e9221813beb3387fa5a9548bf87ae4d6ebf130c1145b652fae" exitCode=0 Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.520245 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xr297" event={"ID":"0dd114e4-8912-4bbd-831c-40d9621154a4","Type":"ContainerDied","Data":"085b545711ae72e9221813beb3387fa5a9548bf87ae4d6ebf130c1145b652fae"} Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.520556 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xr297" event={"ID":"0dd114e4-8912-4bbd-831c-40d9621154a4","Type":"ContainerDied","Data":"5eac150afd1659ccde21b3d8017b5af51f1687367361c1e76ce13ef93e16711b"} Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.520305 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xr297" Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.520584 4685 scope.go:117] "RemoveContainer" containerID="085b545711ae72e9221813beb3387fa5a9548bf87ae4d6ebf130c1145b652fae" Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.542595 4685 scope.go:117] "RemoveContainer" containerID="37900305fe9cd3eadb6aa7d95d08abf435250cba073f81fecd27fa177d6bc667" Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.573117 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xr297"] Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.580933 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xr297"] Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.582614 4685 scope.go:117] "RemoveContainer" containerID="277d20e5d8b05111d50357698cd9c0a4a85c0cabd80cbcded068ecf8a0a9c1c5" Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.638341 4685 scope.go:117] "RemoveContainer" containerID="085b545711ae72e9221813beb3387fa5a9548bf87ae4d6ebf130c1145b652fae" Oct 13 09:30:54 crc kubenswrapper[4685]: E1013 09:30:54.638798 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"085b545711ae72e9221813beb3387fa5a9548bf87ae4d6ebf130c1145b652fae\": container with ID starting with 085b545711ae72e9221813beb3387fa5a9548bf87ae4d6ebf130c1145b652fae not found: ID does not exist" containerID="085b545711ae72e9221813beb3387fa5a9548bf87ae4d6ebf130c1145b652fae" Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.638835 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"085b545711ae72e9221813beb3387fa5a9548bf87ae4d6ebf130c1145b652fae"} err="failed to get container status \"085b545711ae72e9221813beb3387fa5a9548bf87ae4d6ebf130c1145b652fae\": rpc error: code = NotFound desc = could not find container \"085b545711ae72e9221813beb3387fa5a9548bf87ae4d6ebf130c1145b652fae\": container with ID starting with 085b545711ae72e9221813beb3387fa5a9548bf87ae4d6ebf130c1145b652fae not found: ID does not exist" Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.638861 4685 scope.go:117] "RemoveContainer" containerID="37900305fe9cd3eadb6aa7d95d08abf435250cba073f81fecd27fa177d6bc667" Oct 13 09:30:54 crc kubenswrapper[4685]: E1013 09:30:54.639310 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37900305fe9cd3eadb6aa7d95d08abf435250cba073f81fecd27fa177d6bc667\": container with ID starting with 37900305fe9cd3eadb6aa7d95d08abf435250cba073f81fecd27fa177d6bc667 not found: ID does not exist" containerID="37900305fe9cd3eadb6aa7d95d08abf435250cba073f81fecd27fa177d6bc667" Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.639332 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37900305fe9cd3eadb6aa7d95d08abf435250cba073f81fecd27fa177d6bc667"} err="failed to get container status \"37900305fe9cd3eadb6aa7d95d08abf435250cba073f81fecd27fa177d6bc667\": rpc error: code = NotFound desc = could not find container \"37900305fe9cd3eadb6aa7d95d08abf435250cba073f81fecd27fa177d6bc667\": container with ID starting with 37900305fe9cd3eadb6aa7d95d08abf435250cba073f81fecd27fa177d6bc667 not found: ID does not exist" Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.639344 4685 scope.go:117] "RemoveContainer" containerID="277d20e5d8b05111d50357698cd9c0a4a85c0cabd80cbcded068ecf8a0a9c1c5" Oct 13 09:30:54 crc kubenswrapper[4685]: E1013 09:30:54.639550 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"277d20e5d8b05111d50357698cd9c0a4a85c0cabd80cbcded068ecf8a0a9c1c5\": container with ID starting with 277d20e5d8b05111d50357698cd9c0a4a85c0cabd80cbcded068ecf8a0a9c1c5 not found: ID does not exist" containerID="277d20e5d8b05111d50357698cd9c0a4a85c0cabd80cbcded068ecf8a0a9c1c5" Oct 13 09:30:54 crc kubenswrapper[4685]: I1013 09:30:54.639572 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"277d20e5d8b05111d50357698cd9c0a4a85c0cabd80cbcded068ecf8a0a9c1c5"} err="failed to get container status \"277d20e5d8b05111d50357698cd9c0a4a85c0cabd80cbcded068ecf8a0a9c1c5\": rpc error: code = NotFound desc = could not find container \"277d20e5d8b05111d50357698cd9c0a4a85c0cabd80cbcded068ecf8a0a9c1c5\": container with ID starting with 277d20e5d8b05111d50357698cd9c0a4a85c0cabd80cbcded068ecf8a0a9c1c5 not found: ID does not exist" Oct 13 09:30:55 crc kubenswrapper[4685]: I1013 09:30:55.525176 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dd114e4-8912-4bbd-831c-40d9621154a4" path="/var/lib/kubelet/pods/0dd114e4-8912-4bbd-831c-40d9621154a4/volumes" Oct 13 09:30:56 crc kubenswrapper[4685]: I1013 09:30:56.620442 4685 scope.go:117] "RemoveContainer" containerID="c41f5b326a5ed917fc98da2be46741a8baa16dd62bb2a2fb843af8060340137f" Oct 13 09:30:58 crc kubenswrapper[4685]: I1013 09:30:58.503231 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:30:58 crc kubenswrapper[4685]: E1013 09:30:58.503809 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:31:09 crc kubenswrapper[4685]: I1013 09:31:09.503154 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:31:09 crc kubenswrapper[4685]: E1013 09:31:09.503898 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:31:11 crc kubenswrapper[4685]: I1013 09:31:11.737635 4685 generic.go:334] "Generic (PLEG): container finished" podID="e955f1da-bde6-47d5-8cdf-fe0d7182c081" containerID="2b0b224a3cf819d933a4be7286887bbf4e1047de1093960c6d487e658181e619" exitCode=0 Oct 13 09:31:11 crc kubenswrapper[4685]: I1013 09:31:11.737742 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" event={"ID":"e955f1da-bde6-47d5-8cdf-fe0d7182c081","Type":"ContainerDied","Data":"2b0b224a3cf819d933a4be7286887bbf4e1047de1093960c6d487e658181e619"} Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.182524 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.298169 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-inventory\") pod \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.298266 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ceilometer-compute-config-data-1\") pod \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.298474 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-telemetry-combined-ca-bundle\") pod \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.298507 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ceilometer-compute-config-data-0\") pod \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.298534 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhswg\" (UniqueName: \"kubernetes.io/projected/e955f1da-bde6-47d5-8cdf-fe0d7182c081-kube-api-access-jhswg\") pod \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.298614 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ssh-key\") pod \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.298683 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ceilometer-compute-config-data-2\") pod \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\" (UID: \"e955f1da-bde6-47d5-8cdf-fe0d7182c081\") " Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.303795 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e955f1da-bde6-47d5-8cdf-fe0d7182c081-kube-api-access-jhswg" (OuterVolumeSpecName: "kube-api-access-jhswg") pod "e955f1da-bde6-47d5-8cdf-fe0d7182c081" (UID: "e955f1da-bde6-47d5-8cdf-fe0d7182c081"). InnerVolumeSpecName "kube-api-access-jhswg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.320791 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "e955f1da-bde6-47d5-8cdf-fe0d7182c081" (UID: "e955f1da-bde6-47d5-8cdf-fe0d7182c081"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.325978 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "e955f1da-bde6-47d5-8cdf-fe0d7182c081" (UID: "e955f1da-bde6-47d5-8cdf-fe0d7182c081"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.326949 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "e955f1da-bde6-47d5-8cdf-fe0d7182c081" (UID: "e955f1da-bde6-47d5-8cdf-fe0d7182c081"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.328379 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e955f1da-bde6-47d5-8cdf-fe0d7182c081" (UID: "e955f1da-bde6-47d5-8cdf-fe0d7182c081"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.330438 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "e955f1da-bde6-47d5-8cdf-fe0d7182c081" (UID: "e955f1da-bde6-47d5-8cdf-fe0d7182c081"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.338398 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-inventory" (OuterVolumeSpecName: "inventory") pod "e955f1da-bde6-47d5-8cdf-fe0d7182c081" (UID: "e955f1da-bde6-47d5-8cdf-fe0d7182c081"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.400815 4685 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.400849 4685 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.400859 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhswg\" (UniqueName: \"kubernetes.io/projected/e955f1da-bde6-47d5-8cdf-fe0d7182c081-kube-api-access-jhswg\") on node \"crc\" DevicePath \"\"" Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.400868 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.400878 4685 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.400888 4685 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.400898 4685 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e955f1da-bde6-47d5-8cdf-fe0d7182c081-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.759294 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" event={"ID":"e955f1da-bde6-47d5-8cdf-fe0d7182c081","Type":"ContainerDied","Data":"eedc752cdfe985ee81cb7c64fb434a55221020a643614cb1a2a315a10af2f345"} Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.759338 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eedc752cdfe985ee81cb7c64fb434a55221020a643614cb1a2a315a10af2f345" Oct 13 09:31:13 crc kubenswrapper[4685]: I1013 09:31:13.759381 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg" Oct 13 09:31:22 crc kubenswrapper[4685]: I1013 09:31:22.503701 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:31:22 crc kubenswrapper[4685]: E1013 09:31:22.504750 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:31:37 crc kubenswrapper[4685]: I1013 09:31:37.503379 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:31:37 crc kubenswrapper[4685]: E1013 09:31:37.504292 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:31:48 crc kubenswrapper[4685]: I1013 09:31:48.502639 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:31:48 crc kubenswrapper[4685]: E1013 09:31:48.504953 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.333251 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 13 09:31:59 crc kubenswrapper[4685]: E1013 09:31:59.333953 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dd114e4-8912-4bbd-831c-40d9621154a4" containerName="extract-content" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.333967 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dd114e4-8912-4bbd-831c-40d9621154a4" containerName="extract-content" Oct 13 09:31:59 crc kubenswrapper[4685]: E1013 09:31:59.333995 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dd114e4-8912-4bbd-831c-40d9621154a4" containerName="registry-server" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.334001 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dd114e4-8912-4bbd-831c-40d9621154a4" containerName="registry-server" Oct 13 09:31:59 crc kubenswrapper[4685]: E1013 09:31:59.334016 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e955f1da-bde6-47d5-8cdf-fe0d7182c081" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.334023 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="e955f1da-bde6-47d5-8cdf-fe0d7182c081" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 13 09:31:59 crc kubenswrapper[4685]: E1013 09:31:59.334030 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dd114e4-8912-4bbd-831c-40d9621154a4" containerName="extract-utilities" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.334036 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dd114e4-8912-4bbd-831c-40d9621154a4" containerName="extract-utilities" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.334203 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="e955f1da-bde6-47d5-8cdf-fe0d7182c081" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.334222 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dd114e4-8912-4bbd-831c-40d9621154a4" containerName="registry-server" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.335168 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.338740 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.339486 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-pfs5k" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.339739 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.341523 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.346094 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.358763 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9465\" (UniqueName: \"kubernetes.io/projected/804b452c-e552-4715-901c-061f1dc7db41-kube-api-access-g9465\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.359000 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/804b452c-e552-4715-901c-061f1dc7db41-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.359141 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/804b452c-e552-4715-901c-061f1dc7db41-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.359669 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/804b452c-e552-4715-901c-061f1dc7db41-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.359770 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/804b452c-e552-4715-901c-061f1dc7db41-config-data\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.359883 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.359976 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/804b452c-e552-4715-901c-061f1dc7db41-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.360088 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/804b452c-e552-4715-901c-061f1dc7db41-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.360165 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/804b452c-e552-4715-901c-061f1dc7db41-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.461497 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/804b452c-e552-4715-901c-061f1dc7db41-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.461766 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/804b452c-e552-4715-901c-061f1dc7db41-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.461863 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9465\" (UniqueName: \"kubernetes.io/projected/804b452c-e552-4715-901c-061f1dc7db41-kube-api-access-g9465\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.461999 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/804b452c-e552-4715-901c-061f1dc7db41-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.462116 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/804b452c-e552-4715-901c-061f1dc7db41-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.462201 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/804b452c-e552-4715-901c-061f1dc7db41-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.462298 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/804b452c-e552-4715-901c-061f1dc7db41-config-data\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.462445 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.462532 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/804b452c-e552-4715-901c-061f1dc7db41-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.462376 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/804b452c-e552-4715-901c-061f1dc7db41-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.463212 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/804b452c-e552-4715-901c-061f1dc7db41-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.463476 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/804b452c-e552-4715-901c-061f1dc7db41-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.463900 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.464246 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/804b452c-e552-4715-901c-061f1dc7db41-config-data\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.472233 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/804b452c-e552-4715-901c-061f1dc7db41-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.478506 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/804b452c-e552-4715-901c-061f1dc7db41-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.479297 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/804b452c-e552-4715-901c-061f1dc7db41-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.481307 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9465\" (UniqueName: \"kubernetes.io/projected/804b452c-e552-4715-901c-061f1dc7db41-kube-api-access-g9465\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.496115 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " pod="openstack/tempest-tests-tempest" Oct 13 09:31:59 crc kubenswrapper[4685]: I1013 09:31:59.667709 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 13 09:32:00 crc kubenswrapper[4685]: I1013 09:32:00.126698 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 13 09:32:00 crc kubenswrapper[4685]: I1013 09:32:00.251591 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"804b452c-e552-4715-901c-061f1dc7db41","Type":"ContainerStarted","Data":"0b28fb8cca561ba163bb6c38e4c4cc946d2629a71eff76041ac101b03c569bd0"} Oct 13 09:32:03 crc kubenswrapper[4685]: I1013 09:32:03.503779 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:32:03 crc kubenswrapper[4685]: E1013 09:32:03.506288 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:32:18 crc kubenswrapper[4685]: I1013 09:32:18.503072 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:32:18 crc kubenswrapper[4685]: E1013 09:32:18.504007 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:32:22 crc kubenswrapper[4685]: I1013 09:32:22.979684 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:32:22 crc kubenswrapper[4685]: I1013 09:32:22.980338 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:32:24 crc kubenswrapper[4685]: I1013 09:32:24.829925 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zbqmb"] Oct 13 09:32:24 crc kubenswrapper[4685]: I1013 09:32:24.834115 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zbqmb" Oct 13 09:32:24 crc kubenswrapper[4685]: I1013 09:32:24.846043 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zbqmb"] Oct 13 09:32:24 crc kubenswrapper[4685]: I1013 09:32:24.972285 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03a84130-346a-4805-8dfc-43d72f8024e7-utilities\") pod \"community-operators-zbqmb\" (UID: \"03a84130-346a-4805-8dfc-43d72f8024e7\") " pod="openshift-marketplace/community-operators-zbqmb" Oct 13 09:32:24 crc kubenswrapper[4685]: I1013 09:32:24.972416 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp7dr\" (UniqueName: \"kubernetes.io/projected/03a84130-346a-4805-8dfc-43d72f8024e7-kube-api-access-xp7dr\") pod \"community-operators-zbqmb\" (UID: \"03a84130-346a-4805-8dfc-43d72f8024e7\") " pod="openshift-marketplace/community-operators-zbqmb" Oct 13 09:32:24 crc kubenswrapper[4685]: I1013 09:32:24.972481 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03a84130-346a-4805-8dfc-43d72f8024e7-catalog-content\") pod \"community-operators-zbqmb\" (UID: \"03a84130-346a-4805-8dfc-43d72f8024e7\") " pod="openshift-marketplace/community-operators-zbqmb" Oct 13 09:32:25 crc kubenswrapper[4685]: I1013 09:32:25.073727 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp7dr\" (UniqueName: \"kubernetes.io/projected/03a84130-346a-4805-8dfc-43d72f8024e7-kube-api-access-xp7dr\") pod \"community-operators-zbqmb\" (UID: \"03a84130-346a-4805-8dfc-43d72f8024e7\") " pod="openshift-marketplace/community-operators-zbqmb" Oct 13 09:32:25 crc kubenswrapper[4685]: I1013 09:32:25.073793 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03a84130-346a-4805-8dfc-43d72f8024e7-catalog-content\") pod \"community-operators-zbqmb\" (UID: \"03a84130-346a-4805-8dfc-43d72f8024e7\") " pod="openshift-marketplace/community-operators-zbqmb" Oct 13 09:32:25 crc kubenswrapper[4685]: I1013 09:32:25.073873 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03a84130-346a-4805-8dfc-43d72f8024e7-utilities\") pod \"community-operators-zbqmb\" (UID: \"03a84130-346a-4805-8dfc-43d72f8024e7\") " pod="openshift-marketplace/community-operators-zbqmb" Oct 13 09:32:25 crc kubenswrapper[4685]: I1013 09:32:25.074437 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03a84130-346a-4805-8dfc-43d72f8024e7-catalog-content\") pod \"community-operators-zbqmb\" (UID: \"03a84130-346a-4805-8dfc-43d72f8024e7\") " pod="openshift-marketplace/community-operators-zbqmb" Oct 13 09:32:25 crc kubenswrapper[4685]: I1013 09:32:25.074442 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03a84130-346a-4805-8dfc-43d72f8024e7-utilities\") pod \"community-operators-zbqmb\" (UID: \"03a84130-346a-4805-8dfc-43d72f8024e7\") " pod="openshift-marketplace/community-operators-zbqmb" Oct 13 09:32:25 crc kubenswrapper[4685]: I1013 09:32:25.096400 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp7dr\" (UniqueName: \"kubernetes.io/projected/03a84130-346a-4805-8dfc-43d72f8024e7-kube-api-access-xp7dr\") pod \"community-operators-zbqmb\" (UID: \"03a84130-346a-4805-8dfc-43d72f8024e7\") " pod="openshift-marketplace/community-operators-zbqmb" Oct 13 09:32:25 crc kubenswrapper[4685]: I1013 09:32:25.166473 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zbqmb" Oct 13 09:32:31 crc kubenswrapper[4685]: I1013 09:32:31.502655 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:32:31 crc kubenswrapper[4685]: E1013 09:32:31.504203 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:32:41 crc kubenswrapper[4685]: E1013 09:32:41.826745 4685 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 13 09:32:41 crc kubenswrapper[4685]: E1013 09:32:41.829695 4685 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g9465,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(804b452c-e552-4715-901c-061f1dc7db41): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 09:32:41 crc kubenswrapper[4685]: E1013 09:32:41.831206 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="804b452c-e552-4715-901c-061f1dc7db41" Oct 13 09:32:42 crc kubenswrapper[4685]: I1013 09:32:42.169012 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zbqmb"] Oct 13 09:32:42 crc kubenswrapper[4685]: I1013 09:32:42.502622 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:32:42 crc kubenswrapper[4685]: E1013 09:32:42.503249 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:32:42 crc kubenswrapper[4685]: I1013 09:32:42.664656 4685 generic.go:334] "Generic (PLEG): container finished" podID="03a84130-346a-4805-8dfc-43d72f8024e7" containerID="c41103422c9e7eb1dbbc69d3e1fd7271bfb3281a8874fea3a34f26998d26090e" exitCode=0 Oct 13 09:32:42 crc kubenswrapper[4685]: I1013 09:32:42.665220 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbqmb" event={"ID":"03a84130-346a-4805-8dfc-43d72f8024e7","Type":"ContainerDied","Data":"c41103422c9e7eb1dbbc69d3e1fd7271bfb3281a8874fea3a34f26998d26090e"} Oct 13 09:32:42 crc kubenswrapper[4685]: I1013 09:32:42.665258 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbqmb" event={"ID":"03a84130-346a-4805-8dfc-43d72f8024e7","Type":"ContainerStarted","Data":"78b7ee2ba8e8027c067b81d9453f01759c8a0b02711a4845ba9324d69981903b"} Oct 13 09:32:42 crc kubenswrapper[4685]: E1013 09:32:42.666206 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="804b452c-e552-4715-901c-061f1dc7db41" Oct 13 09:32:44 crc kubenswrapper[4685]: I1013 09:32:44.689433 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbqmb" event={"ID":"03a84130-346a-4805-8dfc-43d72f8024e7","Type":"ContainerStarted","Data":"31f9311de35f65deee72f35fbcdc4e05c43a7c791726e4970cf1251a26d09339"} Oct 13 09:32:45 crc kubenswrapper[4685]: I1013 09:32:45.702832 4685 generic.go:334] "Generic (PLEG): container finished" podID="03a84130-346a-4805-8dfc-43d72f8024e7" containerID="31f9311de35f65deee72f35fbcdc4e05c43a7c791726e4970cf1251a26d09339" exitCode=0 Oct 13 09:32:45 crc kubenswrapper[4685]: I1013 09:32:45.702879 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbqmb" event={"ID":"03a84130-346a-4805-8dfc-43d72f8024e7","Type":"ContainerDied","Data":"31f9311de35f65deee72f35fbcdc4e05c43a7c791726e4970cf1251a26d09339"} Oct 13 09:32:46 crc kubenswrapper[4685]: I1013 09:32:46.714218 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbqmb" event={"ID":"03a84130-346a-4805-8dfc-43d72f8024e7","Type":"ContainerStarted","Data":"1f939ad51c988914d520516369e15c6557a9cf14ff4f81fc1b0c199b1ad90896"} Oct 13 09:32:46 crc kubenswrapper[4685]: I1013 09:32:46.734264 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zbqmb" podStartSLOduration=19.289069376 podStartE2EDuration="22.734246453s" podCreationTimestamp="2025-10-13 09:32:24 +0000 UTC" firstStartedPulling="2025-10-13 09:32:42.666798589 +0000 UTC m=+2887.814674350" lastFinishedPulling="2025-10-13 09:32:46.111975666 +0000 UTC m=+2891.259851427" observedRunningTime="2025-10-13 09:32:46.733657267 +0000 UTC m=+2891.881533038" watchObservedRunningTime="2025-10-13 09:32:46.734246453 +0000 UTC m=+2891.882122214" Oct 13 09:32:52 crc kubenswrapper[4685]: I1013 09:32:52.980267 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:32:52 crc kubenswrapper[4685]: I1013 09:32:52.980732 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:32:54 crc kubenswrapper[4685]: I1013 09:32:54.503836 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:32:54 crc kubenswrapper[4685]: E1013 09:32:54.504097 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:32:55 crc kubenswrapper[4685]: I1013 09:32:55.167657 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zbqmb" Oct 13 09:32:55 crc kubenswrapper[4685]: I1013 09:32:55.168207 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zbqmb" Oct 13 09:32:55 crc kubenswrapper[4685]: I1013 09:32:55.243040 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zbqmb" Oct 13 09:32:55 crc kubenswrapper[4685]: I1013 09:32:55.850904 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zbqmb" Oct 13 09:32:55 crc kubenswrapper[4685]: I1013 09:32:55.997851 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 13 09:32:56 crc kubenswrapper[4685]: I1013 09:32:56.012954 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zbqmb"] Oct 13 09:32:57 crc kubenswrapper[4685]: I1013 09:32:57.821847 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zbqmb" podUID="03a84130-346a-4805-8dfc-43d72f8024e7" containerName="registry-server" containerID="cri-o://1f939ad51c988914d520516369e15c6557a9cf14ff4f81fc1b0c199b1ad90896" gracePeriod=2 Oct 13 09:32:57 crc kubenswrapper[4685]: I1013 09:32:57.822719 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"804b452c-e552-4715-901c-061f1dc7db41","Type":"ContainerStarted","Data":"bd075a44233e1671867198e0218a8ff344fe7d80180cd03e16f30f36d2312695"} Oct 13 09:32:57 crc kubenswrapper[4685]: I1013 09:32:57.849749 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.994282228 podStartE2EDuration="59.849728608s" podCreationTimestamp="2025-10-13 09:31:58 +0000 UTC" firstStartedPulling="2025-10-13 09:32:00.139525342 +0000 UTC m=+2845.287401103" lastFinishedPulling="2025-10-13 09:32:55.994971722 +0000 UTC m=+2901.142847483" observedRunningTime="2025-10-13 09:32:57.840812937 +0000 UTC m=+2902.988688718" watchObservedRunningTime="2025-10-13 09:32:57.849728608 +0000 UTC m=+2902.997604369" Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.244360 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zbqmb" Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.327270 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03a84130-346a-4805-8dfc-43d72f8024e7-utilities\") pod \"03a84130-346a-4805-8dfc-43d72f8024e7\" (UID: \"03a84130-346a-4805-8dfc-43d72f8024e7\") " Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.327335 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03a84130-346a-4805-8dfc-43d72f8024e7-catalog-content\") pod \"03a84130-346a-4805-8dfc-43d72f8024e7\" (UID: \"03a84130-346a-4805-8dfc-43d72f8024e7\") " Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.327482 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xp7dr\" (UniqueName: \"kubernetes.io/projected/03a84130-346a-4805-8dfc-43d72f8024e7-kube-api-access-xp7dr\") pod \"03a84130-346a-4805-8dfc-43d72f8024e7\" (UID: \"03a84130-346a-4805-8dfc-43d72f8024e7\") " Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.327986 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03a84130-346a-4805-8dfc-43d72f8024e7-utilities" (OuterVolumeSpecName: "utilities") pod "03a84130-346a-4805-8dfc-43d72f8024e7" (UID: "03a84130-346a-4805-8dfc-43d72f8024e7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.328192 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03a84130-346a-4805-8dfc-43d72f8024e7-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.333473 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03a84130-346a-4805-8dfc-43d72f8024e7-kube-api-access-xp7dr" (OuterVolumeSpecName: "kube-api-access-xp7dr") pod "03a84130-346a-4805-8dfc-43d72f8024e7" (UID: "03a84130-346a-4805-8dfc-43d72f8024e7"). InnerVolumeSpecName "kube-api-access-xp7dr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.378578 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03a84130-346a-4805-8dfc-43d72f8024e7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03a84130-346a-4805-8dfc-43d72f8024e7" (UID: "03a84130-346a-4805-8dfc-43d72f8024e7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.430166 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xp7dr\" (UniqueName: \"kubernetes.io/projected/03a84130-346a-4805-8dfc-43d72f8024e7-kube-api-access-xp7dr\") on node \"crc\" DevicePath \"\"" Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.430209 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03a84130-346a-4805-8dfc-43d72f8024e7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.832378 4685 generic.go:334] "Generic (PLEG): container finished" podID="03a84130-346a-4805-8dfc-43d72f8024e7" containerID="1f939ad51c988914d520516369e15c6557a9cf14ff4f81fc1b0c199b1ad90896" exitCode=0 Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.832473 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zbqmb" Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.832466 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbqmb" event={"ID":"03a84130-346a-4805-8dfc-43d72f8024e7","Type":"ContainerDied","Data":"1f939ad51c988914d520516369e15c6557a9cf14ff4f81fc1b0c199b1ad90896"} Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.832809 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbqmb" event={"ID":"03a84130-346a-4805-8dfc-43d72f8024e7","Type":"ContainerDied","Data":"78b7ee2ba8e8027c067b81d9453f01759c8a0b02711a4845ba9324d69981903b"} Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.832832 4685 scope.go:117] "RemoveContainer" containerID="1f939ad51c988914d520516369e15c6557a9cf14ff4f81fc1b0c199b1ad90896" Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.872474 4685 scope.go:117] "RemoveContainer" containerID="31f9311de35f65deee72f35fbcdc4e05c43a7c791726e4970cf1251a26d09339" Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.894933 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zbqmb"] Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.904603 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zbqmb"] Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.904753 4685 scope.go:117] "RemoveContainer" containerID="c41103422c9e7eb1dbbc69d3e1fd7271bfb3281a8874fea3a34f26998d26090e" Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.950019 4685 scope.go:117] "RemoveContainer" containerID="1f939ad51c988914d520516369e15c6557a9cf14ff4f81fc1b0c199b1ad90896" Oct 13 09:32:58 crc kubenswrapper[4685]: E1013 09:32:58.950450 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f939ad51c988914d520516369e15c6557a9cf14ff4f81fc1b0c199b1ad90896\": container with ID starting with 1f939ad51c988914d520516369e15c6557a9cf14ff4f81fc1b0c199b1ad90896 not found: ID does not exist" containerID="1f939ad51c988914d520516369e15c6557a9cf14ff4f81fc1b0c199b1ad90896" Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.950489 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f939ad51c988914d520516369e15c6557a9cf14ff4f81fc1b0c199b1ad90896"} err="failed to get container status \"1f939ad51c988914d520516369e15c6557a9cf14ff4f81fc1b0c199b1ad90896\": rpc error: code = NotFound desc = could not find container \"1f939ad51c988914d520516369e15c6557a9cf14ff4f81fc1b0c199b1ad90896\": container with ID starting with 1f939ad51c988914d520516369e15c6557a9cf14ff4f81fc1b0c199b1ad90896 not found: ID does not exist" Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.950517 4685 scope.go:117] "RemoveContainer" containerID="31f9311de35f65deee72f35fbcdc4e05c43a7c791726e4970cf1251a26d09339" Oct 13 09:32:58 crc kubenswrapper[4685]: E1013 09:32:58.950737 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31f9311de35f65deee72f35fbcdc4e05c43a7c791726e4970cf1251a26d09339\": container with ID starting with 31f9311de35f65deee72f35fbcdc4e05c43a7c791726e4970cf1251a26d09339 not found: ID does not exist" containerID="31f9311de35f65deee72f35fbcdc4e05c43a7c791726e4970cf1251a26d09339" Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.950754 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31f9311de35f65deee72f35fbcdc4e05c43a7c791726e4970cf1251a26d09339"} err="failed to get container status \"31f9311de35f65deee72f35fbcdc4e05c43a7c791726e4970cf1251a26d09339\": rpc error: code = NotFound desc = could not find container \"31f9311de35f65deee72f35fbcdc4e05c43a7c791726e4970cf1251a26d09339\": container with ID starting with 31f9311de35f65deee72f35fbcdc4e05c43a7c791726e4970cf1251a26d09339 not found: ID does not exist" Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.950773 4685 scope.go:117] "RemoveContainer" containerID="c41103422c9e7eb1dbbc69d3e1fd7271bfb3281a8874fea3a34f26998d26090e" Oct 13 09:32:58 crc kubenswrapper[4685]: E1013 09:32:58.950982 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c41103422c9e7eb1dbbc69d3e1fd7271bfb3281a8874fea3a34f26998d26090e\": container with ID starting with c41103422c9e7eb1dbbc69d3e1fd7271bfb3281a8874fea3a34f26998d26090e not found: ID does not exist" containerID="c41103422c9e7eb1dbbc69d3e1fd7271bfb3281a8874fea3a34f26998d26090e" Oct 13 09:32:58 crc kubenswrapper[4685]: I1013 09:32:58.951002 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c41103422c9e7eb1dbbc69d3e1fd7271bfb3281a8874fea3a34f26998d26090e"} err="failed to get container status \"c41103422c9e7eb1dbbc69d3e1fd7271bfb3281a8874fea3a34f26998d26090e\": rpc error: code = NotFound desc = could not find container \"c41103422c9e7eb1dbbc69d3e1fd7271bfb3281a8874fea3a34f26998d26090e\": container with ID starting with c41103422c9e7eb1dbbc69d3e1fd7271bfb3281a8874fea3a34f26998d26090e not found: ID does not exist" Oct 13 09:32:59 crc kubenswrapper[4685]: I1013 09:32:59.519370 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03a84130-346a-4805-8dfc-43d72f8024e7" path="/var/lib/kubelet/pods/03a84130-346a-4805-8dfc-43d72f8024e7/volumes" Oct 13 09:33:07 crc kubenswrapper[4685]: I1013 09:33:07.505268 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:33:07 crc kubenswrapper[4685]: E1013 09:33:07.506366 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:33:20 crc kubenswrapper[4685]: I1013 09:33:20.502494 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:33:20 crc kubenswrapper[4685]: E1013 09:33:20.504269 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:33:21 crc kubenswrapper[4685]: I1013 09:33:21.555631 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6r8kv"] Oct 13 09:33:21 crc kubenswrapper[4685]: E1013 09:33:21.556429 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03a84130-346a-4805-8dfc-43d72f8024e7" containerName="extract-content" Oct 13 09:33:21 crc kubenswrapper[4685]: I1013 09:33:21.556449 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="03a84130-346a-4805-8dfc-43d72f8024e7" containerName="extract-content" Oct 13 09:33:21 crc kubenswrapper[4685]: E1013 09:33:21.556481 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03a84130-346a-4805-8dfc-43d72f8024e7" containerName="registry-server" Oct 13 09:33:21 crc kubenswrapper[4685]: I1013 09:33:21.556491 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="03a84130-346a-4805-8dfc-43d72f8024e7" containerName="registry-server" Oct 13 09:33:21 crc kubenswrapper[4685]: E1013 09:33:21.556526 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03a84130-346a-4805-8dfc-43d72f8024e7" containerName="extract-utilities" Oct 13 09:33:21 crc kubenswrapper[4685]: I1013 09:33:21.556535 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="03a84130-346a-4805-8dfc-43d72f8024e7" containerName="extract-utilities" Oct 13 09:33:21 crc kubenswrapper[4685]: I1013 09:33:21.556776 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="03a84130-346a-4805-8dfc-43d72f8024e7" containerName="registry-server" Oct 13 09:33:21 crc kubenswrapper[4685]: I1013 09:33:21.558526 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6r8kv" Oct 13 09:33:21 crc kubenswrapper[4685]: I1013 09:33:21.577383 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6r8kv"] Oct 13 09:33:21 crc kubenswrapper[4685]: I1013 09:33:21.682052 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bfe0aab-7097-4527-ab06-1f3f5e5a81dc-utilities\") pod \"redhat-marketplace-6r8kv\" (UID: \"7bfe0aab-7097-4527-ab06-1f3f5e5a81dc\") " pod="openshift-marketplace/redhat-marketplace-6r8kv" Oct 13 09:33:21 crc kubenswrapper[4685]: I1013 09:33:21.682178 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pnl6\" (UniqueName: \"kubernetes.io/projected/7bfe0aab-7097-4527-ab06-1f3f5e5a81dc-kube-api-access-5pnl6\") pod \"redhat-marketplace-6r8kv\" (UID: \"7bfe0aab-7097-4527-ab06-1f3f5e5a81dc\") " pod="openshift-marketplace/redhat-marketplace-6r8kv" Oct 13 09:33:21 crc kubenswrapper[4685]: I1013 09:33:21.682311 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bfe0aab-7097-4527-ab06-1f3f5e5a81dc-catalog-content\") pod \"redhat-marketplace-6r8kv\" (UID: \"7bfe0aab-7097-4527-ab06-1f3f5e5a81dc\") " pod="openshift-marketplace/redhat-marketplace-6r8kv" Oct 13 09:33:21 crc kubenswrapper[4685]: I1013 09:33:21.787576 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bfe0aab-7097-4527-ab06-1f3f5e5a81dc-catalog-content\") pod \"redhat-marketplace-6r8kv\" (UID: \"7bfe0aab-7097-4527-ab06-1f3f5e5a81dc\") " pod="openshift-marketplace/redhat-marketplace-6r8kv" Oct 13 09:33:21 crc kubenswrapper[4685]: I1013 09:33:21.787685 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bfe0aab-7097-4527-ab06-1f3f5e5a81dc-catalog-content\") pod \"redhat-marketplace-6r8kv\" (UID: \"7bfe0aab-7097-4527-ab06-1f3f5e5a81dc\") " pod="openshift-marketplace/redhat-marketplace-6r8kv" Oct 13 09:33:21 crc kubenswrapper[4685]: I1013 09:33:21.788247 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bfe0aab-7097-4527-ab06-1f3f5e5a81dc-utilities\") pod \"redhat-marketplace-6r8kv\" (UID: \"7bfe0aab-7097-4527-ab06-1f3f5e5a81dc\") " pod="openshift-marketplace/redhat-marketplace-6r8kv" Oct 13 09:33:21 crc kubenswrapper[4685]: I1013 09:33:21.788290 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bfe0aab-7097-4527-ab06-1f3f5e5a81dc-utilities\") pod \"redhat-marketplace-6r8kv\" (UID: \"7bfe0aab-7097-4527-ab06-1f3f5e5a81dc\") " pod="openshift-marketplace/redhat-marketplace-6r8kv" Oct 13 09:33:21 crc kubenswrapper[4685]: I1013 09:33:21.788457 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pnl6\" (UniqueName: \"kubernetes.io/projected/7bfe0aab-7097-4527-ab06-1f3f5e5a81dc-kube-api-access-5pnl6\") pod \"redhat-marketplace-6r8kv\" (UID: \"7bfe0aab-7097-4527-ab06-1f3f5e5a81dc\") " pod="openshift-marketplace/redhat-marketplace-6r8kv" Oct 13 09:33:21 crc kubenswrapper[4685]: I1013 09:33:21.825597 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pnl6\" (UniqueName: \"kubernetes.io/projected/7bfe0aab-7097-4527-ab06-1f3f5e5a81dc-kube-api-access-5pnl6\") pod \"redhat-marketplace-6r8kv\" (UID: \"7bfe0aab-7097-4527-ab06-1f3f5e5a81dc\") " pod="openshift-marketplace/redhat-marketplace-6r8kv" Oct 13 09:33:21 crc kubenswrapper[4685]: I1013 09:33:21.937708 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6r8kv" Oct 13 09:33:22 crc kubenswrapper[4685]: I1013 09:33:22.507758 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6r8kv"] Oct 13 09:33:22 crc kubenswrapper[4685]: I1013 09:33:22.980178 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:33:22 crc kubenswrapper[4685]: I1013 09:33:22.980243 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:33:22 crc kubenswrapper[4685]: I1013 09:33:22.980296 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 09:33:22 crc kubenswrapper[4685]: I1013 09:33:22.981042 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f1a9443768e1ae7075cbb2eb7a5ba6388414738c043869b315940a2a75bdea21"} pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 09:33:22 crc kubenswrapper[4685]: I1013 09:33:22.981386 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" containerID="cri-o://f1a9443768e1ae7075cbb2eb7a5ba6388414738c043869b315940a2a75bdea21" gracePeriod=600 Oct 13 09:33:23 crc kubenswrapper[4685]: I1013 09:33:23.084366 4685 generic.go:334] "Generic (PLEG): container finished" podID="7bfe0aab-7097-4527-ab06-1f3f5e5a81dc" containerID="e235848f22b2b33c691e63c77b7be533fd2c624a431ac424e4301808201f8448" exitCode=0 Oct 13 09:33:23 crc kubenswrapper[4685]: I1013 09:33:23.084435 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6r8kv" event={"ID":"7bfe0aab-7097-4527-ab06-1f3f5e5a81dc","Type":"ContainerDied","Data":"e235848f22b2b33c691e63c77b7be533fd2c624a431ac424e4301808201f8448"} Oct 13 09:33:23 crc kubenswrapper[4685]: I1013 09:33:23.084474 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6r8kv" event={"ID":"7bfe0aab-7097-4527-ab06-1f3f5e5a81dc","Type":"ContainerStarted","Data":"5a64dc5021c541554300a3289567dd4f4b4b0cec1f731705e34d6312cdfb8f45"} Oct 13 09:33:23 crc kubenswrapper[4685]: I1013 09:33:23.087046 4685 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 13 09:33:24 crc kubenswrapper[4685]: I1013 09:33:24.113985 4685 generic.go:334] "Generic (PLEG): container finished" podID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerID="f1a9443768e1ae7075cbb2eb7a5ba6388414738c043869b315940a2a75bdea21" exitCode=0 Oct 13 09:33:24 crc kubenswrapper[4685]: I1013 09:33:24.114151 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerDied","Data":"f1a9443768e1ae7075cbb2eb7a5ba6388414738c043869b315940a2a75bdea21"} Oct 13 09:33:24 crc kubenswrapper[4685]: I1013 09:33:24.114344 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerStarted","Data":"dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6"} Oct 13 09:33:24 crc kubenswrapper[4685]: I1013 09:33:24.114366 4685 scope.go:117] "RemoveContainer" containerID="3e36a3f0cee5e73b37d1140c34df167541ec60118cc06be47fe7d2df816805ad" Oct 13 09:33:25 crc kubenswrapper[4685]: I1013 09:33:25.128119 4685 generic.go:334] "Generic (PLEG): container finished" podID="7bfe0aab-7097-4527-ab06-1f3f5e5a81dc" containerID="59ba465354d26b6f3ebbef33108e717075268e8d505797ebcdafa1fb7a067ddf" exitCode=0 Oct 13 09:33:25 crc kubenswrapper[4685]: I1013 09:33:25.128176 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6r8kv" event={"ID":"7bfe0aab-7097-4527-ab06-1f3f5e5a81dc","Type":"ContainerDied","Data":"59ba465354d26b6f3ebbef33108e717075268e8d505797ebcdafa1fb7a067ddf"} Oct 13 09:33:26 crc kubenswrapper[4685]: I1013 09:33:26.144543 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6r8kv" event={"ID":"7bfe0aab-7097-4527-ab06-1f3f5e5a81dc","Type":"ContainerStarted","Data":"5d14b9dc991379859a490735f8e299686edee15b021cef5656749d43b834823c"} Oct 13 09:33:26 crc kubenswrapper[4685]: I1013 09:33:26.168379 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6r8kv" podStartSLOduration=2.6690228190000003 podStartE2EDuration="5.168357226s" podCreationTimestamp="2025-10-13 09:33:21 +0000 UTC" firstStartedPulling="2025-10-13 09:33:23.086369221 +0000 UTC m=+2928.234244982" lastFinishedPulling="2025-10-13 09:33:25.585703628 +0000 UTC m=+2930.733579389" observedRunningTime="2025-10-13 09:33:26.159319571 +0000 UTC m=+2931.307195362" watchObservedRunningTime="2025-10-13 09:33:26.168357226 +0000 UTC m=+2931.316232987" Oct 13 09:33:31 crc kubenswrapper[4685]: I1013 09:33:31.939086 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6r8kv" Oct 13 09:33:31 crc kubenswrapper[4685]: I1013 09:33:31.939585 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6r8kv" Oct 13 09:33:31 crc kubenswrapper[4685]: I1013 09:33:31.996059 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6r8kv" Oct 13 09:33:32 crc kubenswrapper[4685]: I1013 09:33:32.256745 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6r8kv" Oct 13 09:33:32 crc kubenswrapper[4685]: I1013 09:33:32.314539 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6r8kv"] Oct 13 09:33:33 crc kubenswrapper[4685]: I1013 09:33:33.502697 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:33:33 crc kubenswrapper[4685]: E1013 09:33:33.503488 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:33:34 crc kubenswrapper[4685]: I1013 09:33:34.217057 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6r8kv" podUID="7bfe0aab-7097-4527-ab06-1f3f5e5a81dc" containerName="registry-server" containerID="cri-o://5d14b9dc991379859a490735f8e299686edee15b021cef5656749d43b834823c" gracePeriod=2 Oct 13 09:33:34 crc kubenswrapper[4685]: I1013 09:33:34.672440 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6r8kv" Oct 13 09:33:34 crc kubenswrapper[4685]: I1013 09:33:34.822385 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bfe0aab-7097-4527-ab06-1f3f5e5a81dc-catalog-content\") pod \"7bfe0aab-7097-4527-ab06-1f3f5e5a81dc\" (UID: \"7bfe0aab-7097-4527-ab06-1f3f5e5a81dc\") " Oct 13 09:33:34 crc kubenswrapper[4685]: I1013 09:33:34.822425 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pnl6\" (UniqueName: \"kubernetes.io/projected/7bfe0aab-7097-4527-ab06-1f3f5e5a81dc-kube-api-access-5pnl6\") pod \"7bfe0aab-7097-4527-ab06-1f3f5e5a81dc\" (UID: \"7bfe0aab-7097-4527-ab06-1f3f5e5a81dc\") " Oct 13 09:33:34 crc kubenswrapper[4685]: I1013 09:33:34.822446 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bfe0aab-7097-4527-ab06-1f3f5e5a81dc-utilities\") pod \"7bfe0aab-7097-4527-ab06-1f3f5e5a81dc\" (UID: \"7bfe0aab-7097-4527-ab06-1f3f5e5a81dc\") " Oct 13 09:33:34 crc kubenswrapper[4685]: I1013 09:33:34.823458 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bfe0aab-7097-4527-ab06-1f3f5e5a81dc-utilities" (OuterVolumeSpecName: "utilities") pod "7bfe0aab-7097-4527-ab06-1f3f5e5a81dc" (UID: "7bfe0aab-7097-4527-ab06-1f3f5e5a81dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:33:34 crc kubenswrapper[4685]: I1013 09:33:34.828576 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bfe0aab-7097-4527-ab06-1f3f5e5a81dc-kube-api-access-5pnl6" (OuterVolumeSpecName: "kube-api-access-5pnl6") pod "7bfe0aab-7097-4527-ab06-1f3f5e5a81dc" (UID: "7bfe0aab-7097-4527-ab06-1f3f5e5a81dc"). InnerVolumeSpecName "kube-api-access-5pnl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:33:34 crc kubenswrapper[4685]: I1013 09:33:34.837208 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bfe0aab-7097-4527-ab06-1f3f5e5a81dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7bfe0aab-7097-4527-ab06-1f3f5e5a81dc" (UID: "7bfe0aab-7097-4527-ab06-1f3f5e5a81dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:33:34 crc kubenswrapper[4685]: I1013 09:33:34.924701 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bfe0aab-7097-4527-ab06-1f3f5e5a81dc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 09:33:34 crc kubenswrapper[4685]: I1013 09:33:34.924743 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pnl6\" (UniqueName: \"kubernetes.io/projected/7bfe0aab-7097-4527-ab06-1f3f5e5a81dc-kube-api-access-5pnl6\") on node \"crc\" DevicePath \"\"" Oct 13 09:33:34 crc kubenswrapper[4685]: I1013 09:33:34.924757 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bfe0aab-7097-4527-ab06-1f3f5e5a81dc-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 09:33:35 crc kubenswrapper[4685]: I1013 09:33:35.226945 4685 generic.go:334] "Generic (PLEG): container finished" podID="7bfe0aab-7097-4527-ab06-1f3f5e5a81dc" containerID="5d14b9dc991379859a490735f8e299686edee15b021cef5656749d43b834823c" exitCode=0 Oct 13 09:33:35 crc kubenswrapper[4685]: I1013 09:33:35.226991 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6r8kv" event={"ID":"7bfe0aab-7097-4527-ab06-1f3f5e5a81dc","Type":"ContainerDied","Data":"5d14b9dc991379859a490735f8e299686edee15b021cef5656749d43b834823c"} Oct 13 09:33:35 crc kubenswrapper[4685]: I1013 09:33:35.227022 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6r8kv" event={"ID":"7bfe0aab-7097-4527-ab06-1f3f5e5a81dc","Type":"ContainerDied","Data":"5a64dc5021c541554300a3289567dd4f4b4b0cec1f731705e34d6312cdfb8f45"} Oct 13 09:33:35 crc kubenswrapper[4685]: I1013 09:33:35.227041 4685 scope.go:117] "RemoveContainer" containerID="5d14b9dc991379859a490735f8e299686edee15b021cef5656749d43b834823c" Oct 13 09:33:35 crc kubenswrapper[4685]: I1013 09:33:35.227198 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6r8kv" Oct 13 09:33:35 crc kubenswrapper[4685]: I1013 09:33:35.254972 4685 scope.go:117] "RemoveContainer" containerID="59ba465354d26b6f3ebbef33108e717075268e8d505797ebcdafa1fb7a067ddf" Oct 13 09:33:35 crc kubenswrapper[4685]: I1013 09:33:35.265652 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6r8kv"] Oct 13 09:33:35 crc kubenswrapper[4685]: I1013 09:33:35.274900 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6r8kv"] Oct 13 09:33:35 crc kubenswrapper[4685]: I1013 09:33:35.292519 4685 scope.go:117] "RemoveContainer" containerID="e235848f22b2b33c691e63c77b7be533fd2c624a431ac424e4301808201f8448" Oct 13 09:33:35 crc kubenswrapper[4685]: I1013 09:33:35.328201 4685 scope.go:117] "RemoveContainer" containerID="5d14b9dc991379859a490735f8e299686edee15b021cef5656749d43b834823c" Oct 13 09:33:35 crc kubenswrapper[4685]: E1013 09:33:35.328747 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d14b9dc991379859a490735f8e299686edee15b021cef5656749d43b834823c\": container with ID starting with 5d14b9dc991379859a490735f8e299686edee15b021cef5656749d43b834823c not found: ID does not exist" containerID="5d14b9dc991379859a490735f8e299686edee15b021cef5656749d43b834823c" Oct 13 09:33:35 crc kubenswrapper[4685]: I1013 09:33:35.328804 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d14b9dc991379859a490735f8e299686edee15b021cef5656749d43b834823c"} err="failed to get container status \"5d14b9dc991379859a490735f8e299686edee15b021cef5656749d43b834823c\": rpc error: code = NotFound desc = could not find container \"5d14b9dc991379859a490735f8e299686edee15b021cef5656749d43b834823c\": container with ID starting with 5d14b9dc991379859a490735f8e299686edee15b021cef5656749d43b834823c not found: ID does not exist" Oct 13 09:33:35 crc kubenswrapper[4685]: I1013 09:33:35.328839 4685 scope.go:117] "RemoveContainer" containerID="59ba465354d26b6f3ebbef33108e717075268e8d505797ebcdafa1fb7a067ddf" Oct 13 09:33:35 crc kubenswrapper[4685]: E1013 09:33:35.329348 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59ba465354d26b6f3ebbef33108e717075268e8d505797ebcdafa1fb7a067ddf\": container with ID starting with 59ba465354d26b6f3ebbef33108e717075268e8d505797ebcdafa1fb7a067ddf not found: ID does not exist" containerID="59ba465354d26b6f3ebbef33108e717075268e8d505797ebcdafa1fb7a067ddf" Oct 13 09:33:35 crc kubenswrapper[4685]: I1013 09:33:35.329374 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59ba465354d26b6f3ebbef33108e717075268e8d505797ebcdafa1fb7a067ddf"} err="failed to get container status \"59ba465354d26b6f3ebbef33108e717075268e8d505797ebcdafa1fb7a067ddf\": rpc error: code = NotFound desc = could not find container \"59ba465354d26b6f3ebbef33108e717075268e8d505797ebcdafa1fb7a067ddf\": container with ID starting with 59ba465354d26b6f3ebbef33108e717075268e8d505797ebcdafa1fb7a067ddf not found: ID does not exist" Oct 13 09:33:35 crc kubenswrapper[4685]: I1013 09:33:35.329404 4685 scope.go:117] "RemoveContainer" containerID="e235848f22b2b33c691e63c77b7be533fd2c624a431ac424e4301808201f8448" Oct 13 09:33:35 crc kubenswrapper[4685]: E1013 09:33:35.329808 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e235848f22b2b33c691e63c77b7be533fd2c624a431ac424e4301808201f8448\": container with ID starting with e235848f22b2b33c691e63c77b7be533fd2c624a431ac424e4301808201f8448 not found: ID does not exist" containerID="e235848f22b2b33c691e63c77b7be533fd2c624a431ac424e4301808201f8448" Oct 13 09:33:35 crc kubenswrapper[4685]: I1013 09:33:35.329845 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e235848f22b2b33c691e63c77b7be533fd2c624a431ac424e4301808201f8448"} err="failed to get container status \"e235848f22b2b33c691e63c77b7be533fd2c624a431ac424e4301808201f8448\": rpc error: code = NotFound desc = could not find container \"e235848f22b2b33c691e63c77b7be533fd2c624a431ac424e4301808201f8448\": container with ID starting with e235848f22b2b33c691e63c77b7be533fd2c624a431ac424e4301808201f8448 not found: ID does not exist" Oct 13 09:33:35 crc kubenswrapper[4685]: I1013 09:33:35.518549 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bfe0aab-7097-4527-ab06-1f3f5e5a81dc" path="/var/lib/kubelet/pods/7bfe0aab-7097-4527-ab06-1f3f5e5a81dc/volumes" Oct 13 09:33:45 crc kubenswrapper[4685]: I1013 09:33:45.511523 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:33:45 crc kubenswrapper[4685]: E1013 09:33:45.513362 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:33:56 crc kubenswrapper[4685]: I1013 09:33:56.503129 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:33:56 crc kubenswrapper[4685]: E1013 09:33:56.503835 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:34:11 crc kubenswrapper[4685]: I1013 09:34:11.503077 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:34:11 crc kubenswrapper[4685]: E1013 09:34:11.503859 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:34:23 crc kubenswrapper[4685]: I1013 09:34:23.502809 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:34:23 crc kubenswrapper[4685]: E1013 09:34:23.503732 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:34:38 crc kubenswrapper[4685]: I1013 09:34:38.503049 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:34:38 crc kubenswrapper[4685]: E1013 09:34:38.505094 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:34:53 crc kubenswrapper[4685]: I1013 09:34:53.508611 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:34:53 crc kubenswrapper[4685]: E1013 09:34:53.510269 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:35:04 crc kubenswrapper[4685]: I1013 09:35:04.503686 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:35:04 crc kubenswrapper[4685]: E1013 09:35:04.504337 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:35:17 crc kubenswrapper[4685]: I1013 09:35:17.502680 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:35:18 crc kubenswrapper[4685]: I1013 09:35:18.199194 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" event={"ID":"b001d17a-1aea-44ba-86c5-ba6b312156c1","Type":"ContainerStarted","Data":"096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a"} Oct 13 09:35:18 crc kubenswrapper[4685]: I1013 09:35:18.200517 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:35:23 crc kubenswrapper[4685]: I1013 09:35:23.444800 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:35:35 crc kubenswrapper[4685]: I1013 09:35:35.361330 4685 generic.go:334] "Generic (PLEG): container finished" podID="804b452c-e552-4715-901c-061f1dc7db41" containerID="bd075a44233e1671867198e0218a8ff344fe7d80180cd03e16f30f36d2312695" exitCode=0 Oct 13 09:35:35 crc kubenswrapper[4685]: I1013 09:35:35.362104 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"804b452c-e552-4715-901c-061f1dc7db41","Type":"ContainerDied","Data":"bd075a44233e1671867198e0218a8ff344fe7d80180cd03e16f30f36d2312695"} Oct 13 09:35:36 crc kubenswrapper[4685]: I1013 09:35:36.763140 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 13 09:35:36 crc kubenswrapper[4685]: I1013 09:35:36.919037 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/804b452c-e552-4715-901c-061f1dc7db41-openstack-config-secret\") pod \"804b452c-e552-4715-901c-061f1dc7db41\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " Oct 13 09:35:36 crc kubenswrapper[4685]: I1013 09:35:36.919098 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"804b452c-e552-4715-901c-061f1dc7db41\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " Oct 13 09:35:36 crc kubenswrapper[4685]: I1013 09:35:36.919134 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/804b452c-e552-4715-901c-061f1dc7db41-config-data\") pod \"804b452c-e552-4715-901c-061f1dc7db41\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " Oct 13 09:35:36 crc kubenswrapper[4685]: I1013 09:35:36.919199 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9465\" (UniqueName: \"kubernetes.io/projected/804b452c-e552-4715-901c-061f1dc7db41-kube-api-access-g9465\") pod \"804b452c-e552-4715-901c-061f1dc7db41\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " Oct 13 09:35:36 crc kubenswrapper[4685]: I1013 09:35:36.919241 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/804b452c-e552-4715-901c-061f1dc7db41-openstack-config\") pod \"804b452c-e552-4715-901c-061f1dc7db41\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " Oct 13 09:35:36 crc kubenswrapper[4685]: I1013 09:35:36.919283 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/804b452c-e552-4715-901c-061f1dc7db41-test-operator-ephemeral-temporary\") pod \"804b452c-e552-4715-901c-061f1dc7db41\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " Oct 13 09:35:36 crc kubenswrapper[4685]: I1013 09:35:36.919359 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/804b452c-e552-4715-901c-061f1dc7db41-test-operator-ephemeral-workdir\") pod \"804b452c-e552-4715-901c-061f1dc7db41\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " Oct 13 09:35:36 crc kubenswrapper[4685]: I1013 09:35:36.919433 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/804b452c-e552-4715-901c-061f1dc7db41-ssh-key\") pod \"804b452c-e552-4715-901c-061f1dc7db41\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " Oct 13 09:35:36 crc kubenswrapper[4685]: I1013 09:35:36.919544 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/804b452c-e552-4715-901c-061f1dc7db41-ca-certs\") pod \"804b452c-e552-4715-901c-061f1dc7db41\" (UID: \"804b452c-e552-4715-901c-061f1dc7db41\") " Oct 13 09:35:36 crc kubenswrapper[4685]: I1013 09:35:36.920793 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/804b452c-e552-4715-901c-061f1dc7db41-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "804b452c-e552-4715-901c-061f1dc7db41" (UID: "804b452c-e552-4715-901c-061f1dc7db41"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:35:36 crc kubenswrapper[4685]: I1013 09:35:36.921489 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/804b452c-e552-4715-901c-061f1dc7db41-config-data" (OuterVolumeSpecName: "config-data") pod "804b452c-e552-4715-901c-061f1dc7db41" (UID: "804b452c-e552-4715-901c-061f1dc7db41"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:35:36 crc kubenswrapper[4685]: I1013 09:35:36.927637 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/804b452c-e552-4715-901c-061f1dc7db41-kube-api-access-g9465" (OuterVolumeSpecName: "kube-api-access-g9465") pod "804b452c-e552-4715-901c-061f1dc7db41" (UID: "804b452c-e552-4715-901c-061f1dc7db41"). InnerVolumeSpecName "kube-api-access-g9465". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:35:36 crc kubenswrapper[4685]: I1013 09:35:36.930254 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/804b452c-e552-4715-901c-061f1dc7db41-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "804b452c-e552-4715-901c-061f1dc7db41" (UID: "804b452c-e552-4715-901c-061f1dc7db41"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:35:36 crc kubenswrapper[4685]: I1013 09:35:36.930389 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "test-operator-logs") pod "804b452c-e552-4715-901c-061f1dc7db41" (UID: "804b452c-e552-4715-901c-061f1dc7db41"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 13 09:35:36 crc kubenswrapper[4685]: I1013 09:35:36.949313 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/804b452c-e552-4715-901c-061f1dc7db41-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "804b452c-e552-4715-901c-061f1dc7db41" (UID: "804b452c-e552-4715-901c-061f1dc7db41"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:35:36 crc kubenswrapper[4685]: I1013 09:35:36.950645 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/804b452c-e552-4715-901c-061f1dc7db41-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "804b452c-e552-4715-901c-061f1dc7db41" (UID: "804b452c-e552-4715-901c-061f1dc7db41"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:35:36 crc kubenswrapper[4685]: I1013 09:35:36.951881 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/804b452c-e552-4715-901c-061f1dc7db41-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "804b452c-e552-4715-901c-061f1dc7db41" (UID: "804b452c-e552-4715-901c-061f1dc7db41"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:35:36 crc kubenswrapper[4685]: I1013 09:35:36.969340 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/804b452c-e552-4715-901c-061f1dc7db41-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "804b452c-e552-4715-901c-061f1dc7db41" (UID: "804b452c-e552-4715-901c-061f1dc7db41"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:35:37 crc kubenswrapper[4685]: I1013 09:35:37.021319 4685 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/804b452c-e552-4715-901c-061f1dc7db41-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 13 09:35:37 crc kubenswrapper[4685]: I1013 09:35:37.021349 4685 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/804b452c-e552-4715-901c-061f1dc7db41-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 09:35:37 crc kubenswrapper[4685]: I1013 09:35:37.021359 4685 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/804b452c-e552-4715-901c-061f1dc7db41-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 13 09:35:37 crc kubenswrapper[4685]: I1013 09:35:37.021367 4685 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/804b452c-e552-4715-901c-061f1dc7db41-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 13 09:35:37 crc kubenswrapper[4685]: I1013 09:35:37.021401 4685 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 13 09:35:37 crc kubenswrapper[4685]: I1013 09:35:37.021412 4685 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/804b452c-e552-4715-901c-061f1dc7db41-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 09:35:37 crc kubenswrapper[4685]: I1013 09:35:37.021422 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9465\" (UniqueName: \"kubernetes.io/projected/804b452c-e552-4715-901c-061f1dc7db41-kube-api-access-g9465\") on node \"crc\" DevicePath \"\"" Oct 13 09:35:37 crc kubenswrapper[4685]: I1013 09:35:37.021430 4685 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/804b452c-e552-4715-901c-061f1dc7db41-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 13 09:35:37 crc kubenswrapper[4685]: I1013 09:35:37.021439 4685 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/804b452c-e552-4715-901c-061f1dc7db41-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 13 09:35:37 crc kubenswrapper[4685]: I1013 09:35:37.041747 4685 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 13 09:35:37 crc kubenswrapper[4685]: I1013 09:35:37.123550 4685 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 13 09:35:37 crc kubenswrapper[4685]: I1013 09:35:37.382403 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"804b452c-e552-4715-901c-061f1dc7db41","Type":"ContainerDied","Data":"0b28fb8cca561ba163bb6c38e4c4cc946d2629a71eff76041ac101b03c569bd0"} Oct 13 09:35:37 crc kubenswrapper[4685]: I1013 09:35:37.382442 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b28fb8cca561ba163bb6c38e4c4cc946d2629a71eff76041ac101b03c569bd0" Oct 13 09:35:37 crc kubenswrapper[4685]: I1013 09:35:37.382457 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 13 09:35:47 crc kubenswrapper[4685]: I1013 09:35:47.232769 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 13 09:35:47 crc kubenswrapper[4685]: E1013 09:35:47.234746 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bfe0aab-7097-4527-ab06-1f3f5e5a81dc" containerName="registry-server" Oct 13 09:35:47 crc kubenswrapper[4685]: I1013 09:35:47.234848 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bfe0aab-7097-4527-ab06-1f3f5e5a81dc" containerName="registry-server" Oct 13 09:35:47 crc kubenswrapper[4685]: E1013 09:35:47.235073 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bfe0aab-7097-4527-ab06-1f3f5e5a81dc" containerName="extract-content" Oct 13 09:35:47 crc kubenswrapper[4685]: I1013 09:35:47.235163 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bfe0aab-7097-4527-ab06-1f3f5e5a81dc" containerName="extract-content" Oct 13 09:35:47 crc kubenswrapper[4685]: E1013 09:35:47.235302 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bfe0aab-7097-4527-ab06-1f3f5e5a81dc" containerName="extract-utilities" Oct 13 09:35:47 crc kubenswrapper[4685]: I1013 09:35:47.235410 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bfe0aab-7097-4527-ab06-1f3f5e5a81dc" containerName="extract-utilities" Oct 13 09:35:47 crc kubenswrapper[4685]: E1013 09:35:47.235508 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="804b452c-e552-4715-901c-061f1dc7db41" containerName="tempest-tests-tempest-tests-runner" Oct 13 09:35:47 crc kubenswrapper[4685]: I1013 09:35:47.235610 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="804b452c-e552-4715-901c-061f1dc7db41" containerName="tempest-tests-tempest-tests-runner" Oct 13 09:35:47 crc kubenswrapper[4685]: I1013 09:35:47.236529 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bfe0aab-7097-4527-ab06-1f3f5e5a81dc" containerName="registry-server" Oct 13 09:35:47 crc kubenswrapper[4685]: I1013 09:35:47.236635 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="804b452c-e552-4715-901c-061f1dc7db41" containerName="tempest-tests-tempest-tests-runner" Oct 13 09:35:47 crc kubenswrapper[4685]: I1013 09:35:47.237854 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 13 09:35:47 crc kubenswrapper[4685]: I1013 09:35:47.240309 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-pfs5k" Oct 13 09:35:47 crc kubenswrapper[4685]: I1013 09:35:47.247107 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 13 09:35:47 crc kubenswrapper[4685]: I1013 09:35:47.385073 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbkvs\" (UniqueName: \"kubernetes.io/projected/aed7cf9e-6686-44a8-91ef-a1e9613b0c65-kube-api-access-vbkvs\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"aed7cf9e-6686-44a8-91ef-a1e9613b0c65\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 13 09:35:47 crc kubenswrapper[4685]: I1013 09:35:47.385302 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"aed7cf9e-6686-44a8-91ef-a1e9613b0c65\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 13 09:35:47 crc kubenswrapper[4685]: I1013 09:35:47.486613 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"aed7cf9e-6686-44a8-91ef-a1e9613b0c65\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 13 09:35:47 crc kubenswrapper[4685]: I1013 09:35:47.486682 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbkvs\" (UniqueName: \"kubernetes.io/projected/aed7cf9e-6686-44a8-91ef-a1e9613b0c65-kube-api-access-vbkvs\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"aed7cf9e-6686-44a8-91ef-a1e9613b0c65\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 13 09:35:47 crc kubenswrapper[4685]: I1013 09:35:47.486968 4685 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"aed7cf9e-6686-44a8-91ef-a1e9613b0c65\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 13 09:35:47 crc kubenswrapper[4685]: I1013 09:35:47.508259 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbkvs\" (UniqueName: \"kubernetes.io/projected/aed7cf9e-6686-44a8-91ef-a1e9613b0c65-kube-api-access-vbkvs\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"aed7cf9e-6686-44a8-91ef-a1e9613b0c65\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 13 09:35:47 crc kubenswrapper[4685]: I1013 09:35:47.533161 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"aed7cf9e-6686-44a8-91ef-a1e9613b0c65\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 13 09:35:47 crc kubenswrapper[4685]: I1013 09:35:47.580574 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 13 09:35:48 crc kubenswrapper[4685]: I1013 09:35:48.066446 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 13 09:35:48 crc kubenswrapper[4685]: W1013 09:35:48.076992 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaed7cf9e_6686_44a8_91ef_a1e9613b0c65.slice/crio-24b6541d9500f00d84d09e470d8940905012a55eae1e219d8b2bdddde15e7ab4 WatchSource:0}: Error finding container 24b6541d9500f00d84d09e470d8940905012a55eae1e219d8b2bdddde15e7ab4: Status 404 returned error can't find the container with id 24b6541d9500f00d84d09e470d8940905012a55eae1e219d8b2bdddde15e7ab4 Oct 13 09:35:48 crc kubenswrapper[4685]: I1013 09:35:48.571689 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"aed7cf9e-6686-44a8-91ef-a1e9613b0c65","Type":"ContainerStarted","Data":"24b6541d9500f00d84d09e470d8940905012a55eae1e219d8b2bdddde15e7ab4"} Oct 13 09:35:50 crc kubenswrapper[4685]: I1013 09:35:50.594129 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"aed7cf9e-6686-44a8-91ef-a1e9613b0c65","Type":"ContainerStarted","Data":"014fb3079c18df76dbca7579a38817d205ffab25d0db5ddb347101e01293e097"} Oct 13 09:35:50 crc kubenswrapper[4685]: I1013 09:35:50.615982 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.952914674 podStartE2EDuration="3.615965866s" podCreationTimestamp="2025-10-13 09:35:47 +0000 UTC" firstStartedPulling="2025-10-13 09:35:48.080393241 +0000 UTC m=+3073.228269002" lastFinishedPulling="2025-10-13 09:35:49.743444433 +0000 UTC m=+3074.891320194" observedRunningTime="2025-10-13 09:35:50.607445396 +0000 UTC m=+3075.755321157" watchObservedRunningTime="2025-10-13 09:35:50.615965866 +0000 UTC m=+3075.763841647" Oct 13 09:35:52 crc kubenswrapper[4685]: I1013 09:35:52.980121 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:35:52 crc kubenswrapper[4685]: I1013 09:35:52.980477 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:36:06 crc kubenswrapper[4685]: I1013 09:36:06.475386 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-cv7wl/must-gather-vjvgz"] Oct 13 09:36:06 crc kubenswrapper[4685]: I1013 09:36:06.483168 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cv7wl/must-gather-vjvgz" Oct 13 09:36:06 crc kubenswrapper[4685]: I1013 09:36:06.491047 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-cv7wl/must-gather-vjvgz"] Oct 13 09:36:06 crc kubenswrapper[4685]: I1013 09:36:06.494829 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-cv7wl"/"openshift-service-ca.crt" Oct 13 09:36:06 crc kubenswrapper[4685]: I1013 09:36:06.495288 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-cv7wl"/"kube-root-ca.crt" Oct 13 09:36:06 crc kubenswrapper[4685]: I1013 09:36:06.602837 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/60144059-a904-431c-bf75-12f923f281f8-must-gather-output\") pod \"must-gather-vjvgz\" (UID: \"60144059-a904-431c-bf75-12f923f281f8\") " pod="openshift-must-gather-cv7wl/must-gather-vjvgz" Oct 13 09:36:06 crc kubenswrapper[4685]: I1013 09:36:06.602939 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bn5fp\" (UniqueName: \"kubernetes.io/projected/60144059-a904-431c-bf75-12f923f281f8-kube-api-access-bn5fp\") pod \"must-gather-vjvgz\" (UID: \"60144059-a904-431c-bf75-12f923f281f8\") " pod="openshift-must-gather-cv7wl/must-gather-vjvgz" Oct 13 09:36:06 crc kubenswrapper[4685]: I1013 09:36:06.704093 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/60144059-a904-431c-bf75-12f923f281f8-must-gather-output\") pod \"must-gather-vjvgz\" (UID: \"60144059-a904-431c-bf75-12f923f281f8\") " pod="openshift-must-gather-cv7wl/must-gather-vjvgz" Oct 13 09:36:06 crc kubenswrapper[4685]: I1013 09:36:06.704149 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bn5fp\" (UniqueName: \"kubernetes.io/projected/60144059-a904-431c-bf75-12f923f281f8-kube-api-access-bn5fp\") pod \"must-gather-vjvgz\" (UID: \"60144059-a904-431c-bf75-12f923f281f8\") " pod="openshift-must-gather-cv7wl/must-gather-vjvgz" Oct 13 09:36:06 crc kubenswrapper[4685]: I1013 09:36:06.704726 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/60144059-a904-431c-bf75-12f923f281f8-must-gather-output\") pod \"must-gather-vjvgz\" (UID: \"60144059-a904-431c-bf75-12f923f281f8\") " pod="openshift-must-gather-cv7wl/must-gather-vjvgz" Oct 13 09:36:06 crc kubenswrapper[4685]: I1013 09:36:06.721792 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bn5fp\" (UniqueName: \"kubernetes.io/projected/60144059-a904-431c-bf75-12f923f281f8-kube-api-access-bn5fp\") pod \"must-gather-vjvgz\" (UID: \"60144059-a904-431c-bf75-12f923f281f8\") " pod="openshift-must-gather-cv7wl/must-gather-vjvgz" Oct 13 09:36:06 crc kubenswrapper[4685]: I1013 09:36:06.804784 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cv7wl/must-gather-vjvgz" Oct 13 09:36:07 crc kubenswrapper[4685]: W1013 09:36:07.450099 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60144059_a904_431c_bf75_12f923f281f8.slice/crio-9c98e58662a7ae78cec9e7ccc868e6ed7f8e28d79a6aef5b42375c51406b2eab WatchSource:0}: Error finding container 9c98e58662a7ae78cec9e7ccc868e6ed7f8e28d79a6aef5b42375c51406b2eab: Status 404 returned error can't find the container with id 9c98e58662a7ae78cec9e7ccc868e6ed7f8e28d79a6aef5b42375c51406b2eab Oct 13 09:36:07 crc kubenswrapper[4685]: I1013 09:36:07.450880 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-cv7wl/must-gather-vjvgz"] Oct 13 09:36:07 crc kubenswrapper[4685]: I1013 09:36:07.768821 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cv7wl/must-gather-vjvgz" event={"ID":"60144059-a904-431c-bf75-12f923f281f8","Type":"ContainerStarted","Data":"9c98e58662a7ae78cec9e7ccc868e6ed7f8e28d79a6aef5b42375c51406b2eab"} Oct 13 09:36:14 crc kubenswrapper[4685]: I1013 09:36:14.837459 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cv7wl/must-gather-vjvgz" event={"ID":"60144059-a904-431c-bf75-12f923f281f8","Type":"ContainerStarted","Data":"e456529df4d346d29295c8c35f85e2d7edb3c8ebf1e431f09349bccd91d5996b"} Oct 13 09:36:15 crc kubenswrapper[4685]: I1013 09:36:15.849755 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cv7wl/must-gather-vjvgz" event={"ID":"60144059-a904-431c-bf75-12f923f281f8","Type":"ContainerStarted","Data":"a839318753cc0e31e5b657f58bd63b2f3948800873127f682fb52b95d77e2eb9"} Oct 13 09:36:15 crc kubenswrapper[4685]: I1013 09:36:15.878506 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-cv7wl/must-gather-vjvgz" podStartSLOduration=3.058407119 podStartE2EDuration="9.878489176s" podCreationTimestamp="2025-10-13 09:36:06 +0000 UTC" firstStartedPulling="2025-10-13 09:36:07.451887023 +0000 UTC m=+3092.599762784" lastFinishedPulling="2025-10-13 09:36:14.27196908 +0000 UTC m=+3099.419844841" observedRunningTime="2025-10-13 09:36:15.870578513 +0000 UTC m=+3101.018454294" watchObservedRunningTime="2025-10-13 09:36:15.878489176 +0000 UTC m=+3101.026364937" Oct 13 09:36:18 crc kubenswrapper[4685]: I1013 09:36:18.447672 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-cv7wl/crc-debug-m5dqd"] Oct 13 09:36:18 crc kubenswrapper[4685]: I1013 09:36:18.450130 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cv7wl/crc-debug-m5dqd" Oct 13 09:36:18 crc kubenswrapper[4685]: I1013 09:36:18.453590 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-cv7wl"/"default-dockercfg-5ndcb" Oct 13 09:36:18 crc kubenswrapper[4685]: I1013 09:36:18.539762 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/89a961c3-dae7-49bd-baa6-f7d465af0c3e-host\") pod \"crc-debug-m5dqd\" (UID: \"89a961c3-dae7-49bd-baa6-f7d465af0c3e\") " pod="openshift-must-gather-cv7wl/crc-debug-m5dqd" Oct 13 09:36:18 crc kubenswrapper[4685]: I1013 09:36:18.539840 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd98x\" (UniqueName: \"kubernetes.io/projected/89a961c3-dae7-49bd-baa6-f7d465af0c3e-kube-api-access-xd98x\") pod \"crc-debug-m5dqd\" (UID: \"89a961c3-dae7-49bd-baa6-f7d465af0c3e\") " pod="openshift-must-gather-cv7wl/crc-debug-m5dqd" Oct 13 09:36:18 crc kubenswrapper[4685]: I1013 09:36:18.641968 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/89a961c3-dae7-49bd-baa6-f7d465af0c3e-host\") pod \"crc-debug-m5dqd\" (UID: \"89a961c3-dae7-49bd-baa6-f7d465af0c3e\") " pod="openshift-must-gather-cv7wl/crc-debug-m5dqd" Oct 13 09:36:18 crc kubenswrapper[4685]: I1013 09:36:18.642068 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd98x\" (UniqueName: \"kubernetes.io/projected/89a961c3-dae7-49bd-baa6-f7d465af0c3e-kube-api-access-xd98x\") pod \"crc-debug-m5dqd\" (UID: \"89a961c3-dae7-49bd-baa6-f7d465af0c3e\") " pod="openshift-must-gather-cv7wl/crc-debug-m5dqd" Oct 13 09:36:18 crc kubenswrapper[4685]: I1013 09:36:18.642582 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/89a961c3-dae7-49bd-baa6-f7d465af0c3e-host\") pod \"crc-debug-m5dqd\" (UID: \"89a961c3-dae7-49bd-baa6-f7d465af0c3e\") " pod="openshift-must-gather-cv7wl/crc-debug-m5dqd" Oct 13 09:36:18 crc kubenswrapper[4685]: I1013 09:36:18.685803 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd98x\" (UniqueName: \"kubernetes.io/projected/89a961c3-dae7-49bd-baa6-f7d465af0c3e-kube-api-access-xd98x\") pod \"crc-debug-m5dqd\" (UID: \"89a961c3-dae7-49bd-baa6-f7d465af0c3e\") " pod="openshift-must-gather-cv7wl/crc-debug-m5dqd" Oct 13 09:36:18 crc kubenswrapper[4685]: I1013 09:36:18.769192 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cv7wl/crc-debug-m5dqd" Oct 13 09:36:18 crc kubenswrapper[4685]: I1013 09:36:18.873393 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cv7wl/crc-debug-m5dqd" event={"ID":"89a961c3-dae7-49bd-baa6-f7d465af0c3e","Type":"ContainerStarted","Data":"c09bc7036ce3922de79e264d724c9a26aaea08c60e054a44ff704564ee1896ed"} Oct 13 09:36:22 crc kubenswrapper[4685]: I1013 09:36:22.979810 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:36:22 crc kubenswrapper[4685]: I1013 09:36:22.980261 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:36:32 crc kubenswrapper[4685]: I1013 09:36:32.004629 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cv7wl/crc-debug-m5dqd" event={"ID":"89a961c3-dae7-49bd-baa6-f7d465af0c3e","Type":"ContainerStarted","Data":"5b7df5bcacf6cf5af9ee108468642b8914d6b54fff86e3fd325685c9e1c15b10"} Oct 13 09:36:32 crc kubenswrapper[4685]: I1013 09:36:32.021311 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-cv7wl/crc-debug-m5dqd" podStartSLOduration=1.948107129 podStartE2EDuration="14.021295425s" podCreationTimestamp="2025-10-13 09:36:18 +0000 UTC" firstStartedPulling="2025-10-13 09:36:18.836136003 +0000 UTC m=+3103.984011764" lastFinishedPulling="2025-10-13 09:36:30.909324299 +0000 UTC m=+3116.057200060" observedRunningTime="2025-10-13 09:36:32.015580871 +0000 UTC m=+3117.163456632" watchObservedRunningTime="2025-10-13 09:36:32.021295425 +0000 UTC m=+3117.169171186" Oct 13 09:36:52 crc kubenswrapper[4685]: I1013 09:36:52.979726 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:36:52 crc kubenswrapper[4685]: I1013 09:36:52.980361 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:36:52 crc kubenswrapper[4685]: I1013 09:36:52.980417 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 09:36:52 crc kubenswrapper[4685]: I1013 09:36:52.981375 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6"} pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 09:36:52 crc kubenswrapper[4685]: I1013 09:36:52.981443 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" containerID="cri-o://dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" gracePeriod=600 Oct 13 09:36:53 crc kubenswrapper[4685]: E1013 09:36:53.115648 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:36:53 crc kubenswrapper[4685]: I1013 09:36:53.208549 4685 generic.go:334] "Generic (PLEG): container finished" podID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" exitCode=0 Oct 13 09:36:53 crc kubenswrapper[4685]: I1013 09:36:53.208592 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerDied","Data":"dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6"} Oct 13 09:36:53 crc kubenswrapper[4685]: I1013 09:36:53.208626 4685 scope.go:117] "RemoveContainer" containerID="f1a9443768e1ae7075cbb2eb7a5ba6388414738c043869b315940a2a75bdea21" Oct 13 09:36:53 crc kubenswrapper[4685]: I1013 09:36:53.209626 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:36:53 crc kubenswrapper[4685]: E1013 09:36:53.209956 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:37:04 crc kubenswrapper[4685]: I1013 09:37:04.502868 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:37:04 crc kubenswrapper[4685]: E1013 09:37:04.503614 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:37:12 crc kubenswrapper[4685]: I1013 09:37:12.362113 4685 generic.go:334] "Generic (PLEG): container finished" podID="89a961c3-dae7-49bd-baa6-f7d465af0c3e" containerID="5b7df5bcacf6cf5af9ee108468642b8914d6b54fff86e3fd325685c9e1c15b10" exitCode=0 Oct 13 09:37:12 crc kubenswrapper[4685]: I1013 09:37:12.362711 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cv7wl/crc-debug-m5dqd" event={"ID":"89a961c3-dae7-49bd-baa6-f7d465af0c3e","Type":"ContainerDied","Data":"5b7df5bcacf6cf5af9ee108468642b8914d6b54fff86e3fd325685c9e1c15b10"} Oct 13 09:37:13 crc kubenswrapper[4685]: I1013 09:37:13.548416 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cv7wl/crc-debug-m5dqd" Oct 13 09:37:13 crc kubenswrapper[4685]: I1013 09:37:13.566735 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xd98x\" (UniqueName: \"kubernetes.io/projected/89a961c3-dae7-49bd-baa6-f7d465af0c3e-kube-api-access-xd98x\") pod \"89a961c3-dae7-49bd-baa6-f7d465af0c3e\" (UID: \"89a961c3-dae7-49bd-baa6-f7d465af0c3e\") " Oct 13 09:37:13 crc kubenswrapper[4685]: I1013 09:37:13.566885 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/89a961c3-dae7-49bd-baa6-f7d465af0c3e-host\") pod \"89a961c3-dae7-49bd-baa6-f7d465af0c3e\" (UID: \"89a961c3-dae7-49bd-baa6-f7d465af0c3e\") " Oct 13 09:37:13 crc kubenswrapper[4685]: I1013 09:37:13.566987 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/89a961c3-dae7-49bd-baa6-f7d465af0c3e-host" (OuterVolumeSpecName: "host") pod "89a961c3-dae7-49bd-baa6-f7d465af0c3e" (UID: "89a961c3-dae7-49bd-baa6-f7d465af0c3e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 09:37:13 crc kubenswrapper[4685]: I1013 09:37:13.568622 4685 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/89a961c3-dae7-49bd-baa6-f7d465af0c3e-host\") on node \"crc\" DevicePath \"\"" Oct 13 09:37:13 crc kubenswrapper[4685]: I1013 09:37:13.575524 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89a961c3-dae7-49bd-baa6-f7d465af0c3e-kube-api-access-xd98x" (OuterVolumeSpecName: "kube-api-access-xd98x") pod "89a961c3-dae7-49bd-baa6-f7d465af0c3e" (UID: "89a961c3-dae7-49bd-baa6-f7d465af0c3e"). InnerVolumeSpecName "kube-api-access-xd98x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:37:13 crc kubenswrapper[4685]: I1013 09:37:13.588219 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-cv7wl/crc-debug-m5dqd"] Oct 13 09:37:13 crc kubenswrapper[4685]: I1013 09:37:13.595648 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-cv7wl/crc-debug-m5dqd"] Oct 13 09:37:13 crc kubenswrapper[4685]: I1013 09:37:13.670524 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xd98x\" (UniqueName: \"kubernetes.io/projected/89a961c3-dae7-49bd-baa6-f7d465af0c3e-kube-api-access-xd98x\") on node \"crc\" DevicePath \"\"" Oct 13 09:37:14 crc kubenswrapper[4685]: I1013 09:37:14.430162 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c09bc7036ce3922de79e264d724c9a26aaea08c60e054a44ff704564ee1896ed" Oct 13 09:37:14 crc kubenswrapper[4685]: I1013 09:37:14.430219 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cv7wl/crc-debug-m5dqd" Oct 13 09:37:14 crc kubenswrapper[4685]: I1013 09:37:14.833127 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-cv7wl/crc-debug-jb2gm"] Oct 13 09:37:14 crc kubenswrapper[4685]: E1013 09:37:14.833566 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89a961c3-dae7-49bd-baa6-f7d465af0c3e" containerName="container-00" Oct 13 09:37:14 crc kubenswrapper[4685]: I1013 09:37:14.833578 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="89a961c3-dae7-49bd-baa6-f7d465af0c3e" containerName="container-00" Oct 13 09:37:14 crc kubenswrapper[4685]: I1013 09:37:14.833770 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="89a961c3-dae7-49bd-baa6-f7d465af0c3e" containerName="container-00" Oct 13 09:37:14 crc kubenswrapper[4685]: I1013 09:37:14.834378 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cv7wl/crc-debug-jb2gm" Oct 13 09:37:14 crc kubenswrapper[4685]: I1013 09:37:14.836262 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-cv7wl"/"default-dockercfg-5ndcb" Oct 13 09:37:14 crc kubenswrapper[4685]: I1013 09:37:14.895574 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f521ce10-ac81-41bf-9bf2-48bb722ce0fc-host\") pod \"crc-debug-jb2gm\" (UID: \"f521ce10-ac81-41bf-9bf2-48bb722ce0fc\") " pod="openshift-must-gather-cv7wl/crc-debug-jb2gm" Oct 13 09:37:14 crc kubenswrapper[4685]: I1013 09:37:14.895674 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghrbm\" (UniqueName: \"kubernetes.io/projected/f521ce10-ac81-41bf-9bf2-48bb722ce0fc-kube-api-access-ghrbm\") pod \"crc-debug-jb2gm\" (UID: \"f521ce10-ac81-41bf-9bf2-48bb722ce0fc\") " pod="openshift-must-gather-cv7wl/crc-debug-jb2gm" Oct 13 09:37:14 crc kubenswrapper[4685]: I1013 09:37:14.997333 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f521ce10-ac81-41bf-9bf2-48bb722ce0fc-host\") pod \"crc-debug-jb2gm\" (UID: \"f521ce10-ac81-41bf-9bf2-48bb722ce0fc\") " pod="openshift-must-gather-cv7wl/crc-debug-jb2gm" Oct 13 09:37:14 crc kubenswrapper[4685]: I1013 09:37:14.997669 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghrbm\" (UniqueName: \"kubernetes.io/projected/f521ce10-ac81-41bf-9bf2-48bb722ce0fc-kube-api-access-ghrbm\") pod \"crc-debug-jb2gm\" (UID: \"f521ce10-ac81-41bf-9bf2-48bb722ce0fc\") " pod="openshift-must-gather-cv7wl/crc-debug-jb2gm" Oct 13 09:37:14 crc kubenswrapper[4685]: I1013 09:37:14.997521 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f521ce10-ac81-41bf-9bf2-48bb722ce0fc-host\") pod \"crc-debug-jb2gm\" (UID: \"f521ce10-ac81-41bf-9bf2-48bb722ce0fc\") " pod="openshift-must-gather-cv7wl/crc-debug-jb2gm" Oct 13 09:37:15 crc kubenswrapper[4685]: I1013 09:37:15.016858 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghrbm\" (UniqueName: \"kubernetes.io/projected/f521ce10-ac81-41bf-9bf2-48bb722ce0fc-kube-api-access-ghrbm\") pod \"crc-debug-jb2gm\" (UID: \"f521ce10-ac81-41bf-9bf2-48bb722ce0fc\") " pod="openshift-must-gather-cv7wl/crc-debug-jb2gm" Oct 13 09:37:15 crc kubenswrapper[4685]: I1013 09:37:15.155580 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cv7wl/crc-debug-jb2gm" Oct 13 09:37:15 crc kubenswrapper[4685]: I1013 09:37:15.439212 4685 generic.go:334] "Generic (PLEG): container finished" podID="f521ce10-ac81-41bf-9bf2-48bb722ce0fc" containerID="9de1401b38adcb6703c5ce010ce28506a4368a5fd5bbce0a69be23faedc86029" exitCode=0 Oct 13 09:37:15 crc kubenswrapper[4685]: I1013 09:37:15.439515 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cv7wl/crc-debug-jb2gm" event={"ID":"f521ce10-ac81-41bf-9bf2-48bb722ce0fc","Type":"ContainerDied","Data":"9de1401b38adcb6703c5ce010ce28506a4368a5fd5bbce0a69be23faedc86029"} Oct 13 09:37:15 crc kubenswrapper[4685]: I1013 09:37:15.439544 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cv7wl/crc-debug-jb2gm" event={"ID":"f521ce10-ac81-41bf-9bf2-48bb722ce0fc","Type":"ContainerStarted","Data":"1072519962e59973431e674fb0c324079bdd4a4b29c940f8a7d7b718ffb5c11b"} Oct 13 09:37:15 crc kubenswrapper[4685]: I1013 09:37:15.520553 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89a961c3-dae7-49bd-baa6-f7d465af0c3e" path="/var/lib/kubelet/pods/89a961c3-dae7-49bd-baa6-f7d465af0c3e/volumes" Oct 13 09:37:15 crc kubenswrapper[4685]: I1013 09:37:15.853531 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-cv7wl/crc-debug-jb2gm"] Oct 13 09:37:15 crc kubenswrapper[4685]: I1013 09:37:15.899483 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-cv7wl/crc-debug-jb2gm"] Oct 13 09:37:16 crc kubenswrapper[4685]: I1013 09:37:16.543406 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cv7wl/crc-debug-jb2gm" Oct 13 09:37:16 crc kubenswrapper[4685]: I1013 09:37:16.622363 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghrbm\" (UniqueName: \"kubernetes.io/projected/f521ce10-ac81-41bf-9bf2-48bb722ce0fc-kube-api-access-ghrbm\") pod \"f521ce10-ac81-41bf-9bf2-48bb722ce0fc\" (UID: \"f521ce10-ac81-41bf-9bf2-48bb722ce0fc\") " Oct 13 09:37:16 crc kubenswrapper[4685]: I1013 09:37:16.622506 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f521ce10-ac81-41bf-9bf2-48bb722ce0fc-host\") pod \"f521ce10-ac81-41bf-9bf2-48bb722ce0fc\" (UID: \"f521ce10-ac81-41bf-9bf2-48bb722ce0fc\") " Oct 13 09:37:16 crc kubenswrapper[4685]: I1013 09:37:16.622559 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f521ce10-ac81-41bf-9bf2-48bb722ce0fc-host" (OuterVolumeSpecName: "host") pod "f521ce10-ac81-41bf-9bf2-48bb722ce0fc" (UID: "f521ce10-ac81-41bf-9bf2-48bb722ce0fc"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 09:37:16 crc kubenswrapper[4685]: I1013 09:37:16.623271 4685 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f521ce10-ac81-41bf-9bf2-48bb722ce0fc-host\") on node \"crc\" DevicePath \"\"" Oct 13 09:37:16 crc kubenswrapper[4685]: I1013 09:37:16.642337 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f521ce10-ac81-41bf-9bf2-48bb722ce0fc-kube-api-access-ghrbm" (OuterVolumeSpecName: "kube-api-access-ghrbm") pod "f521ce10-ac81-41bf-9bf2-48bb722ce0fc" (UID: "f521ce10-ac81-41bf-9bf2-48bb722ce0fc"). InnerVolumeSpecName "kube-api-access-ghrbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:37:16 crc kubenswrapper[4685]: I1013 09:37:16.725266 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghrbm\" (UniqueName: \"kubernetes.io/projected/f521ce10-ac81-41bf-9bf2-48bb722ce0fc-kube-api-access-ghrbm\") on node \"crc\" DevicePath \"\"" Oct 13 09:37:17 crc kubenswrapper[4685]: I1013 09:37:17.131435 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-cv7wl/crc-debug-7g6mt"] Oct 13 09:37:17 crc kubenswrapper[4685]: E1013 09:37:17.131790 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f521ce10-ac81-41bf-9bf2-48bb722ce0fc" containerName="container-00" Oct 13 09:37:17 crc kubenswrapper[4685]: I1013 09:37:17.131802 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="f521ce10-ac81-41bf-9bf2-48bb722ce0fc" containerName="container-00" Oct 13 09:37:17 crc kubenswrapper[4685]: I1013 09:37:17.132004 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="f521ce10-ac81-41bf-9bf2-48bb722ce0fc" containerName="container-00" Oct 13 09:37:17 crc kubenswrapper[4685]: I1013 09:37:17.132598 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cv7wl/crc-debug-7g6mt" Oct 13 09:37:17 crc kubenswrapper[4685]: I1013 09:37:17.233170 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmj4n\" (UniqueName: \"kubernetes.io/projected/cadad5a8-4707-484a-9320-7e4f7875b96f-kube-api-access-rmj4n\") pod \"crc-debug-7g6mt\" (UID: \"cadad5a8-4707-484a-9320-7e4f7875b96f\") " pod="openshift-must-gather-cv7wl/crc-debug-7g6mt" Oct 13 09:37:17 crc kubenswrapper[4685]: I1013 09:37:17.233255 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cadad5a8-4707-484a-9320-7e4f7875b96f-host\") pod \"crc-debug-7g6mt\" (UID: \"cadad5a8-4707-484a-9320-7e4f7875b96f\") " pod="openshift-must-gather-cv7wl/crc-debug-7g6mt" Oct 13 09:37:17 crc kubenswrapper[4685]: I1013 09:37:17.335615 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmj4n\" (UniqueName: \"kubernetes.io/projected/cadad5a8-4707-484a-9320-7e4f7875b96f-kube-api-access-rmj4n\") pod \"crc-debug-7g6mt\" (UID: \"cadad5a8-4707-484a-9320-7e4f7875b96f\") " pod="openshift-must-gather-cv7wl/crc-debug-7g6mt" Oct 13 09:37:17 crc kubenswrapper[4685]: I1013 09:37:17.336026 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cadad5a8-4707-484a-9320-7e4f7875b96f-host\") pod \"crc-debug-7g6mt\" (UID: \"cadad5a8-4707-484a-9320-7e4f7875b96f\") " pod="openshift-must-gather-cv7wl/crc-debug-7g6mt" Oct 13 09:37:17 crc kubenswrapper[4685]: I1013 09:37:17.336126 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cadad5a8-4707-484a-9320-7e4f7875b96f-host\") pod \"crc-debug-7g6mt\" (UID: \"cadad5a8-4707-484a-9320-7e4f7875b96f\") " pod="openshift-must-gather-cv7wl/crc-debug-7g6mt" Oct 13 09:37:17 crc kubenswrapper[4685]: I1013 09:37:17.362681 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmj4n\" (UniqueName: \"kubernetes.io/projected/cadad5a8-4707-484a-9320-7e4f7875b96f-kube-api-access-rmj4n\") pod \"crc-debug-7g6mt\" (UID: \"cadad5a8-4707-484a-9320-7e4f7875b96f\") " pod="openshift-must-gather-cv7wl/crc-debug-7g6mt" Oct 13 09:37:17 crc kubenswrapper[4685]: I1013 09:37:17.456386 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cv7wl/crc-debug-7g6mt" Oct 13 09:37:17 crc kubenswrapper[4685]: I1013 09:37:17.459181 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1072519962e59973431e674fb0c324079bdd4a4b29c940f8a7d7b718ffb5c11b" Oct 13 09:37:17 crc kubenswrapper[4685]: I1013 09:37:17.459220 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cv7wl/crc-debug-jb2gm" Oct 13 09:37:17 crc kubenswrapper[4685]: I1013 09:37:17.504607 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:37:17 crc kubenswrapper[4685]: E1013 09:37:17.504818 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:37:17 crc kubenswrapper[4685]: I1013 09:37:17.516431 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f521ce10-ac81-41bf-9bf2-48bb722ce0fc" path="/var/lib/kubelet/pods/f521ce10-ac81-41bf-9bf2-48bb722ce0fc/volumes" Oct 13 09:37:18 crc kubenswrapper[4685]: I1013 09:37:18.469542 4685 generic.go:334] "Generic (PLEG): container finished" podID="cadad5a8-4707-484a-9320-7e4f7875b96f" containerID="833fa61178811564773f1325feba73a3052e9124b2410d4851094055ee189753" exitCode=0 Oct 13 09:37:18 crc kubenswrapper[4685]: I1013 09:37:18.469643 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cv7wl/crc-debug-7g6mt" event={"ID":"cadad5a8-4707-484a-9320-7e4f7875b96f","Type":"ContainerDied","Data":"833fa61178811564773f1325feba73a3052e9124b2410d4851094055ee189753"} Oct 13 09:37:18 crc kubenswrapper[4685]: I1013 09:37:18.469842 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cv7wl/crc-debug-7g6mt" event={"ID":"cadad5a8-4707-484a-9320-7e4f7875b96f","Type":"ContainerStarted","Data":"416bc53a107b4edea8e3bf7dab068214b3d397840c2856dabd3ba4daebd91878"} Oct 13 09:37:18 crc kubenswrapper[4685]: I1013 09:37:18.503189 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-cv7wl/crc-debug-7g6mt"] Oct 13 09:37:18 crc kubenswrapper[4685]: I1013 09:37:18.515565 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-cv7wl/crc-debug-7g6mt"] Oct 13 09:37:19 crc kubenswrapper[4685]: I1013 09:37:19.607442 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cv7wl/crc-debug-7g6mt" Oct 13 09:37:19 crc kubenswrapper[4685]: I1013 09:37:19.677995 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmj4n\" (UniqueName: \"kubernetes.io/projected/cadad5a8-4707-484a-9320-7e4f7875b96f-kube-api-access-rmj4n\") pod \"cadad5a8-4707-484a-9320-7e4f7875b96f\" (UID: \"cadad5a8-4707-484a-9320-7e4f7875b96f\") " Oct 13 09:37:19 crc kubenswrapper[4685]: I1013 09:37:19.678204 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cadad5a8-4707-484a-9320-7e4f7875b96f-host\") pod \"cadad5a8-4707-484a-9320-7e4f7875b96f\" (UID: \"cadad5a8-4707-484a-9320-7e4f7875b96f\") " Oct 13 09:37:19 crc kubenswrapper[4685]: I1013 09:37:19.678734 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cadad5a8-4707-484a-9320-7e4f7875b96f-host" (OuterVolumeSpecName: "host") pod "cadad5a8-4707-484a-9320-7e4f7875b96f" (UID: "cadad5a8-4707-484a-9320-7e4f7875b96f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 09:37:19 crc kubenswrapper[4685]: I1013 09:37:19.683203 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cadad5a8-4707-484a-9320-7e4f7875b96f-kube-api-access-rmj4n" (OuterVolumeSpecName: "kube-api-access-rmj4n") pod "cadad5a8-4707-484a-9320-7e4f7875b96f" (UID: "cadad5a8-4707-484a-9320-7e4f7875b96f"). InnerVolumeSpecName "kube-api-access-rmj4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:37:19 crc kubenswrapper[4685]: I1013 09:37:19.780562 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmj4n\" (UniqueName: \"kubernetes.io/projected/cadad5a8-4707-484a-9320-7e4f7875b96f-kube-api-access-rmj4n\") on node \"crc\" DevicePath \"\"" Oct 13 09:37:19 crc kubenswrapper[4685]: I1013 09:37:19.780601 4685 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cadad5a8-4707-484a-9320-7e4f7875b96f-host\") on node \"crc\" DevicePath \"\"" Oct 13 09:37:20 crc kubenswrapper[4685]: I1013 09:37:20.487279 4685 scope.go:117] "RemoveContainer" containerID="833fa61178811564773f1325feba73a3052e9124b2410d4851094055ee189753" Oct 13 09:37:20 crc kubenswrapper[4685]: I1013 09:37:20.487307 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cv7wl/crc-debug-7g6mt" Oct 13 09:37:21 crc kubenswrapper[4685]: I1013 09:37:21.513845 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cadad5a8-4707-484a-9320-7e4f7875b96f" path="/var/lib/kubelet/pods/cadad5a8-4707-484a-9320-7e4f7875b96f/volumes" Oct 13 09:37:27 crc kubenswrapper[4685]: I1013 09:37:27.838406 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7cc9c86456-5hsl7_27d72d0b-473c-4983-be92-7b3eca7d5c45/barbican-api/0.log" Oct 13 09:37:27 crc kubenswrapper[4685]: I1013 09:37:27.915420 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7cc9c86456-5hsl7_27d72d0b-473c-4983-be92-7b3eca7d5c45/barbican-api-log/0.log" Oct 13 09:37:28 crc kubenswrapper[4685]: I1013 09:37:28.029369 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-58b69dbf78-cznmg_ad40cbfc-60f4-4ff0-9106-90a9941d5c10/barbican-keystone-listener/0.log" Oct 13 09:37:28 crc kubenswrapper[4685]: I1013 09:37:28.272265 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-f7b8c695-sw5hg_102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11/barbican-worker/0.log" Oct 13 09:37:28 crc kubenswrapper[4685]: I1013 09:37:28.307725 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-58b69dbf78-cznmg_ad40cbfc-60f4-4ff0-9106-90a9941d5c10/barbican-keystone-listener-log/0.log" Oct 13 09:37:28 crc kubenswrapper[4685]: I1013 09:37:28.321017 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-f7b8c695-sw5hg_102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11/barbican-worker-log/0.log" Oct 13 09:37:28 crc kubenswrapper[4685]: I1013 09:37:28.488147 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl_5056abd4-630c-49fe-9ddf-1294f0f6b55a/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:37:28 crc kubenswrapper[4685]: I1013 09:37:28.536025 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5aa884f8-807f-407b-9ea0-ea0812fcd1a2/ceilometer-central-agent/0.log" Oct 13 09:37:28 crc kubenswrapper[4685]: I1013 09:37:28.702241 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5aa884f8-807f-407b-9ea0-ea0812fcd1a2/ceilometer-notification-agent/0.log" Oct 13 09:37:28 crc kubenswrapper[4685]: I1013 09:37:28.747237 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5aa884f8-807f-407b-9ea0-ea0812fcd1a2/proxy-httpd/0.log" Oct 13 09:37:28 crc kubenswrapper[4685]: I1013 09:37:28.749249 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5aa884f8-807f-407b-9ea0-ea0812fcd1a2/sg-core/0.log" Oct 13 09:37:28 crc kubenswrapper[4685]: I1013 09:37:28.935065 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f/cinder-api/0.log" Oct 13 09:37:28 crc kubenswrapper[4685]: I1013 09:37:28.952949 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f/cinder-api-log/0.log" Oct 13 09:37:29 crc kubenswrapper[4685]: I1013 09:37:29.070674 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c60c26e9-8d95-47ec-9211-7053c42c471b/cinder-scheduler/0.log" Oct 13 09:37:29 crc kubenswrapper[4685]: I1013 09:37:29.132252 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c60c26e9-8d95-47ec-9211-7053c42c471b/probe/0.log" Oct 13 09:37:29 crc kubenswrapper[4685]: I1013 09:37:29.273725 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs_f5d774ea-039f-4303-b889-8a6c79077453/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:37:29 crc kubenswrapper[4685]: I1013 09:37:29.397492 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-jhmms_2a07788a-cfaa-4c89-91ec-96999f31a6fd/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:37:29 crc kubenswrapper[4685]: I1013 09:37:29.502736 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:37:29 crc kubenswrapper[4685]: E1013 09:37:29.503340 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:37:29 crc kubenswrapper[4685]: I1013 09:37:29.520907 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-jk24b_c9f2aa20-f6e6-47e7-9244-36184ce2cf1c/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:37:29 crc kubenswrapper[4685]: I1013 09:37:29.618932 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b865b64bc-nsk84_908f2660-8000-46f8-8119-e43b00f79abb/init/0.log" Oct 13 09:37:29 crc kubenswrapper[4685]: I1013 09:37:29.832450 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b865b64bc-nsk84_908f2660-8000-46f8-8119-e43b00f79abb/init/0.log" Oct 13 09:37:29 crc kubenswrapper[4685]: I1013 09:37:29.933330 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc_49edf7b8-3834-40c0-a038-e6c2388842da/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:37:29 crc kubenswrapper[4685]: I1013 09:37:29.953617 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b865b64bc-nsk84_908f2660-8000-46f8-8119-e43b00f79abb/dnsmasq-dns/0.log" Oct 13 09:37:30 crc kubenswrapper[4685]: I1013 09:37:30.141518 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_cf7cb050-76b4-4ee0-b71f-3ad58070fdc3/glance-httpd/0.log" Oct 13 09:37:30 crc kubenswrapper[4685]: I1013 09:37:30.211997 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_cf7cb050-76b4-4ee0-b71f-3ad58070fdc3/glance-log/0.log" Oct 13 09:37:30 crc kubenswrapper[4685]: I1013 09:37:30.354101 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_96ceb8f4-f3db-4f87-b5bf-27fd14076c1b/glance-log/0.log" Oct 13 09:37:30 crc kubenswrapper[4685]: I1013 09:37:30.363657 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_96ceb8f4-f3db-4f87-b5bf-27fd14076c1b/glance-httpd/0.log" Oct 13 09:37:30 crc kubenswrapper[4685]: I1013 09:37:30.567307 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-dbcf576b-msgft_46272601-40bd-43ec-b7a8-6824281972e0/horizon/0.log" Oct 13 09:37:30 crc kubenswrapper[4685]: I1013 09:37:30.872438 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-sc94f_e8ad7931-be11-41ff-9ca5-1ad434026e44/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:37:30 crc kubenswrapper[4685]: I1013 09:37:30.904210 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-dbcf576b-msgft_46272601-40bd-43ec-b7a8-6824281972e0/horizon-log/0.log" Oct 13 09:37:31 crc kubenswrapper[4685]: I1013 09:37:31.054793 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-s2r7z_5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:37:31 crc kubenswrapper[4685]: I1013 09:37:31.323279 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_6a895be7-8da1-4ba2-982a-bde7d7544581/kube-state-metrics/0.log" Oct 13 09:37:31 crc kubenswrapper[4685]: I1013 09:37:31.386190 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-79b7b9958d-7s7c7_869b9e65-4e24-4046-9b64-c1ad8d970be4/keystone-api/0.log" Oct 13 09:37:31 crc kubenswrapper[4685]: I1013 09:37:31.573316 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2_1a2c2127-ae84-45bc-a364-caa9fe26133a/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:37:31 crc kubenswrapper[4685]: I1013 09:37:31.907061 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-786f745fcf-5snx2_4a0c8086-2865-4107-b424-b7820834b297/neutron-api/0.log" Oct 13 09:37:31 crc kubenswrapper[4685]: I1013 09:37:31.947092 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-786f745fcf-5snx2_4a0c8086-2865-4107-b424-b7820834b297/neutron-httpd/0.log" Oct 13 09:37:32 crc kubenswrapper[4685]: I1013 09:37:32.257286 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5_0b2a22b2-f314-42af-95f0-f6213098f535/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:37:32 crc kubenswrapper[4685]: I1013 09:37:32.598691 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_27a5deb0-f57d-4780-9426-7ff101dc91ed/nova-api-log/0.log" Oct 13 09:37:32 crc kubenswrapper[4685]: I1013 09:37:32.621774 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_27a5deb0-f57d-4780-9426-7ff101dc91ed/nova-api-api/0.log" Oct 13 09:37:32 crc kubenswrapper[4685]: I1013 09:37:32.658998 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_95136327-c720-4660-9e60-adb85bd5afa2/nova-cell0-conductor-conductor/0.log" Oct 13 09:37:32 crc kubenswrapper[4685]: I1013 09:37:32.876264 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_b4c0d942-9fdc-4f28-833c-1332cbe23c56/nova-cell1-conductor-conductor/0.log" Oct 13 09:37:33 crc kubenswrapper[4685]: I1013 09:37:33.197148 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_1039fca1-44b0-46d3-8905-daa12aadea65/nova-cell1-novncproxy-novncproxy/0.log" Oct 13 09:37:33 crc kubenswrapper[4685]: I1013 09:37:33.229618 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-f2kmw_5e5b7b28-ee18-4b85-af9c-030cab7da493/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:37:33 crc kubenswrapper[4685]: I1013 09:37:33.620442 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7ddd9484-88bb-4550-b9e7-399e001cfa29/nova-metadata-log/0.log" Oct 13 09:37:33 crc kubenswrapper[4685]: I1013 09:37:33.712496 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_c01b8af0-dd0a-4f1d-9e09-1e6ab0cf7779/nova-scheduler-scheduler/0.log" Oct 13 09:37:34 crc kubenswrapper[4685]: I1013 09:37:34.147357 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20/mysql-bootstrap/0.log" Oct 13 09:37:34 crc kubenswrapper[4685]: I1013 09:37:34.373422 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20/galera/0.log" Oct 13 09:37:34 crc kubenswrapper[4685]: I1013 09:37:34.449953 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20/mysql-bootstrap/0.log" Oct 13 09:37:34 crc kubenswrapper[4685]: I1013 09:37:34.482764 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7ddd9484-88bb-4550-b9e7-399e001cfa29/nova-metadata-metadata/0.log" Oct 13 09:37:34 crc kubenswrapper[4685]: I1013 09:37:34.608716 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f5482688-fe73-4196-ae4d-a306a1c31545/mysql-bootstrap/0.log" Oct 13 09:37:34 crc kubenswrapper[4685]: I1013 09:37:34.844069 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f5482688-fe73-4196-ae4d-a306a1c31545/mysql-bootstrap/0.log" Oct 13 09:37:34 crc kubenswrapper[4685]: I1013 09:37:34.933778 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_31c9baa6-8272-45f1-b8d4-6c89f710fe20/openstackclient/0.log" Oct 13 09:37:34 crc kubenswrapper[4685]: I1013 09:37:34.953596 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f5482688-fe73-4196-ae4d-a306a1c31545/galera/0.log" Oct 13 09:37:35 crc kubenswrapper[4685]: I1013 09:37:35.182780 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-t646z_85cf8998-48c3-49b9-9a68-d5765b785df4/openstack-network-exporter/0.log" Oct 13 09:37:35 crc kubenswrapper[4685]: I1013 09:37:35.305059 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gh7ls_fb7ad71b-d500-457d-81d4-831ba362a8b4/ovsdb-server-init/0.log" Oct 13 09:37:35 crc kubenswrapper[4685]: I1013 09:37:35.621648 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gh7ls_fb7ad71b-d500-457d-81d4-831ba362a8b4/ovsdb-server/0.log" Oct 13 09:37:35 crc kubenswrapper[4685]: I1013 09:37:35.623293 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gh7ls_fb7ad71b-d500-457d-81d4-831ba362a8b4/ovsdb-server-init/0.log" Oct 13 09:37:35 crc kubenswrapper[4685]: I1013 09:37:35.698104 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gh7ls_fb7ad71b-d500-457d-81d4-831ba362a8b4/ovs-vswitchd/0.log" Oct 13 09:37:35 crc kubenswrapper[4685]: I1013 09:37:35.890652 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-tb2c4_9dccff31-26fc-43c7-a679-01173d2f34a9/ovn-controller/0.log" Oct 13 09:37:36 crc kubenswrapper[4685]: I1013 09:37:36.082046 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-nwnvn_57483ac7-dfa4-4538-8605-11b9215432dc/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:37:36 crc kubenswrapper[4685]: I1013 09:37:36.113391 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_108cacab-202f-48f0-91f1-664e68d89e28/openstack-network-exporter/0.log" Oct 13 09:37:36 crc kubenswrapper[4685]: I1013 09:37:36.215958 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_108cacab-202f-48f0-91f1-664e68d89e28/ovn-northd/0.log" Oct 13 09:37:36 crc kubenswrapper[4685]: I1013 09:37:36.380274 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_7ffc3833-03fd-40ef-b247-4f1c512c2e8b/openstack-network-exporter/0.log" Oct 13 09:37:36 crc kubenswrapper[4685]: I1013 09:37:36.451056 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_7ffc3833-03fd-40ef-b247-4f1c512c2e8b/ovsdbserver-nb/0.log" Oct 13 09:37:36 crc kubenswrapper[4685]: I1013 09:37:36.684965 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_2a3319a1-2f17-43cd-9df1-0697ba10aff8/ovsdbserver-sb/0.log" Oct 13 09:37:36 crc kubenswrapper[4685]: I1013 09:37:36.696350 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_2a3319a1-2f17-43cd-9df1-0697ba10aff8/openstack-network-exporter/0.log" Oct 13 09:37:36 crc kubenswrapper[4685]: I1013 09:37:36.878599 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-55b747894d-xd2hx_9a194ffb-9cf0-4167-9c5b-c51bd79c42d7/placement-api/0.log" Oct 13 09:37:37 crc kubenswrapper[4685]: I1013 09:37:37.028509 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-55b747894d-xd2hx_9a194ffb-9cf0-4167-9c5b-c51bd79c42d7/placement-log/0.log" Oct 13 09:37:37 crc kubenswrapper[4685]: I1013 09:37:37.149744 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5f68bc3a-ebea-44a6-9b00-048e6afd1d09/setup-container/0.log" Oct 13 09:37:37 crc kubenswrapper[4685]: I1013 09:37:37.382496 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5f68bc3a-ebea-44a6-9b00-048e6afd1d09/rabbitmq/0.log" Oct 13 09:37:37 crc kubenswrapper[4685]: I1013 09:37:37.433553 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5f68bc3a-ebea-44a6-9b00-048e6afd1d09/setup-container/0.log" Oct 13 09:37:37 crc kubenswrapper[4685]: I1013 09:37:37.517221 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_27720755-e830-4eb4-b0e0-b5dfe9ceb253/setup-container/0.log" Oct 13 09:37:37 crc kubenswrapper[4685]: I1013 09:37:37.780500 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_27720755-e830-4eb4-b0e0-b5dfe9ceb253/rabbitmq/0.log" Oct 13 09:37:37 crc kubenswrapper[4685]: I1013 09:37:37.857147 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_27720755-e830-4eb4-b0e0-b5dfe9ceb253/setup-container/0.log" Oct 13 09:37:37 crc kubenswrapper[4685]: I1013 09:37:37.878381 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj_f6c7cd6b-3a10-490e-af5e-68a812d4b6f6/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:37:38 crc kubenswrapper[4685]: I1013 09:37:38.146139 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-gbjgj_cc857d68-fe9c-4e34-86c7-e26aca5432e6/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:37:38 crc kubenswrapper[4685]: I1013 09:37:38.190937 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc_81fecbf5-ba69-4c64-b5c1-ae9f5e07d881/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:37:38 crc kubenswrapper[4685]: I1013 09:37:38.389120 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-p2spf_82f29acf-d74b-434e-9aaf-7324cbc6c2d0/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:37:38 crc kubenswrapper[4685]: I1013 09:37:38.550711 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-xsmbj_73114199-2f57-451c-8b19-81ad0fbb98f5/ssh-known-hosts-edpm-deployment/0.log" Oct 13 09:37:38 crc kubenswrapper[4685]: I1013 09:37:38.791865 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-55864b6df5-xmtb6_86c3d3b4-cf99-46c5-b238-0efd9798f870/proxy-server/0.log" Oct 13 09:37:38 crc kubenswrapper[4685]: I1013 09:37:38.843795 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-55864b6df5-xmtb6_86c3d3b4-cf99-46c5-b238-0efd9798f870/proxy-httpd/0.log" Oct 13 09:37:39 crc kubenswrapper[4685]: I1013 09:37:39.059928 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-2qrrt_b884b6de-f048-4f14-b8b0-4775fa1d4cd1/swift-ring-rebalance/0.log" Oct 13 09:37:39 crc kubenswrapper[4685]: I1013 09:37:39.116459 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/account-auditor/0.log" Oct 13 09:37:39 crc kubenswrapper[4685]: I1013 09:37:39.144007 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/account-reaper/0.log" Oct 13 09:37:39 crc kubenswrapper[4685]: I1013 09:37:39.389518 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/account-replicator/0.log" Oct 13 09:37:39 crc kubenswrapper[4685]: I1013 09:37:39.398521 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/account-server/0.log" Oct 13 09:37:39 crc kubenswrapper[4685]: I1013 09:37:39.447762 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/container-replicator/0.log" Oct 13 09:37:39 crc kubenswrapper[4685]: I1013 09:37:39.501670 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/container-auditor/0.log" Oct 13 09:37:39 crc kubenswrapper[4685]: I1013 09:37:39.680526 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/object-auditor/0.log" Oct 13 09:37:39 crc kubenswrapper[4685]: I1013 09:37:39.694347 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/container-server/0.log" Oct 13 09:37:39 crc kubenswrapper[4685]: I1013 09:37:39.803142 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/container-updater/0.log" Oct 13 09:37:39 crc kubenswrapper[4685]: I1013 09:37:39.809625 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/object-expirer/0.log" Oct 13 09:37:39 crc kubenswrapper[4685]: I1013 09:37:39.975781 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/object-server/0.log" Oct 13 09:37:39 crc kubenswrapper[4685]: I1013 09:37:39.995821 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/object-replicator/0.log" Oct 13 09:37:40 crc kubenswrapper[4685]: I1013 09:37:40.110195 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/rsync/0.log" Oct 13 09:37:40 crc kubenswrapper[4685]: I1013 09:37:40.156640 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/object-updater/0.log" Oct 13 09:37:40 crc kubenswrapper[4685]: I1013 09:37:40.261419 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/swift-recon-cron/0.log" Oct 13 09:37:40 crc kubenswrapper[4685]: I1013 09:37:40.502245 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:37:40 crc kubenswrapper[4685]: E1013 09:37:40.502505 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:37:40 crc kubenswrapper[4685]: I1013 09:37:40.538228 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg_e955f1da-bde6-47d5-8cdf-fe0d7182c081/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:37:40 crc kubenswrapper[4685]: I1013 09:37:40.623658 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_804b452c-e552-4715-901c-061f1dc7db41/tempest-tests-tempest-tests-runner/0.log" Oct 13 09:37:40 crc kubenswrapper[4685]: I1013 09:37:40.852173 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_aed7cf9e-6686-44a8-91ef-a1e9613b0c65/test-operator-logs-container/0.log" Oct 13 09:37:40 crc kubenswrapper[4685]: I1013 09:37:40.915512 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz_6535c66b-9311-4170-8fa5-c3d79b5cd7af/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:37:51 crc kubenswrapper[4685]: I1013 09:37:51.502489 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:37:51 crc kubenswrapper[4685]: E1013 09:37:51.503262 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:37:52 crc kubenswrapper[4685]: I1013 09:37:52.824735 4685 generic.go:334] "Generic (PLEG): container finished" podID="b001d17a-1aea-44ba-86c5-ba6b312156c1" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" exitCode=1 Oct 13 09:37:52 crc kubenswrapper[4685]: I1013 09:37:52.826775 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" event={"ID":"b001d17a-1aea-44ba-86c5-ba6b312156c1","Type":"ContainerDied","Data":"096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a"} Oct 13 09:37:52 crc kubenswrapper[4685]: I1013 09:37:52.826953 4685 scope.go:117] "RemoveContainer" containerID="69939738a0b31edb19506b2dd29628bd379e4f0c9a35b8c582f14043bb00983b" Oct 13 09:37:52 crc kubenswrapper[4685]: I1013 09:37:52.827819 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:37:52 crc kubenswrapper[4685]: E1013 09:37:52.828231 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:37:53 crc kubenswrapper[4685]: I1013 09:37:53.443774 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:37:53 crc kubenswrapper[4685]: I1013 09:37:53.838183 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:37:53 crc kubenswrapper[4685]: E1013 09:37:53.838510 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:37:55 crc kubenswrapper[4685]: I1013 09:37:55.796256 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_fb5d643a-96c6-4fca-904d-3b4a71e9630c/memcached/0.log" Oct 13 09:38:03 crc kubenswrapper[4685]: I1013 09:38:03.442816 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:38:03 crc kubenswrapper[4685]: I1013 09:38:03.444869 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:38:03 crc kubenswrapper[4685]: E1013 09:38:03.445237 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:38:05 crc kubenswrapper[4685]: I1013 09:38:05.508829 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:38:05 crc kubenswrapper[4685]: E1013 09:38:05.509360 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:38:08 crc kubenswrapper[4685]: I1013 09:38:08.787266 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56_6c967b8a-ec86-413b-a25e-b81f3e9ac2dc/util/0.log" Oct 13 09:38:09 crc kubenswrapper[4685]: I1013 09:38:09.052112 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56_6c967b8a-ec86-413b-a25e-b81f3e9ac2dc/util/0.log" Oct 13 09:38:09 crc kubenswrapper[4685]: I1013 09:38:09.058828 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56_6c967b8a-ec86-413b-a25e-b81f3e9ac2dc/pull/0.log" Oct 13 09:38:09 crc kubenswrapper[4685]: I1013 09:38:09.086706 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56_6c967b8a-ec86-413b-a25e-b81f3e9ac2dc/pull/0.log" Oct 13 09:38:09 crc kubenswrapper[4685]: I1013 09:38:09.227335 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56_6c967b8a-ec86-413b-a25e-b81f3e9ac2dc/util/0.log" Oct 13 09:38:09 crc kubenswrapper[4685]: I1013 09:38:09.307751 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56_6c967b8a-ec86-413b-a25e-b81f3e9ac2dc/pull/0.log" Oct 13 09:38:09 crc kubenswrapper[4685]: I1013 09:38:09.308813 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56_6c967b8a-ec86-413b-a25e-b81f3e9ac2dc/extract/0.log" Oct 13 09:38:09 crc kubenswrapper[4685]: I1013 09:38:09.440343 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-5d7zf_e509e801-67de-4a55-bd22-cf3f73deca81/kube-rbac-proxy/0.log" Oct 13 09:38:09 crc kubenswrapper[4685]: I1013 09:38:09.603704 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-5d7zf_e509e801-67de-4a55-bd22-cf3f73deca81/manager/0.log" Oct 13 09:38:09 crc kubenswrapper[4685]: I1013 09:38:09.612474 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-qbl58_c92c1fba-b02b-4b6c-9570-b75ee60c5e86/kube-rbac-proxy/0.log" Oct 13 09:38:09 crc kubenswrapper[4685]: I1013 09:38:09.730187 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-qbl58_c92c1fba-b02b-4b6c-9570-b75ee60c5e86/manager/0.log" Oct 13 09:38:09 crc kubenswrapper[4685]: I1013 09:38:09.841999 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-hbnzh_92a042d7-669e-48d2-8d7e-8a8da4fc01eb/kube-rbac-proxy/0.log" Oct 13 09:38:09 crc kubenswrapper[4685]: I1013 09:38:09.887561 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-hbnzh_92a042d7-669e-48d2-8d7e-8a8da4fc01eb/manager/0.log" Oct 13 09:38:10 crc kubenswrapper[4685]: I1013 09:38:10.078118 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-zm9bd_54405c7c-61f7-41ea-ae0a-29128b51326c/kube-rbac-proxy/0.log" Oct 13 09:38:10 crc kubenswrapper[4685]: I1013 09:38:10.233465 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-zm9bd_54405c7c-61f7-41ea-ae0a-29128b51326c/manager/0.log" Oct 13 09:38:10 crc kubenswrapper[4685]: I1013 09:38:10.262501 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-mj7x8_6b61f732-3f39-459e-bec0-a6bdd5added4/kube-rbac-proxy/0.log" Oct 13 09:38:10 crc kubenswrapper[4685]: I1013 09:38:10.347282 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-mj7x8_6b61f732-3f39-459e-bec0-a6bdd5added4/manager/0.log" Oct 13 09:38:10 crc kubenswrapper[4685]: I1013 09:38:10.428397 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-sl5lv_eb08316b-7fd3-4d65-88e1-bbc91efcb7c7/kube-rbac-proxy/0.log" Oct 13 09:38:10 crc kubenswrapper[4685]: I1013 09:38:10.559636 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-sl5lv_eb08316b-7fd3-4d65-88e1-bbc91efcb7c7/manager/0.log" Oct 13 09:38:10 crc kubenswrapper[4685]: I1013 09:38:10.671408 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-2bcss_a3e83b7f-5017-44e7-b507-f46a2d3f5488/kube-rbac-proxy/0.log" Oct 13 09:38:10 crc kubenswrapper[4685]: I1013 09:38:10.823828 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-2bcss_a3e83b7f-5017-44e7-b507-f46a2d3f5488/manager/0.log" Oct 13 09:38:10 crc kubenswrapper[4685]: I1013 09:38:10.868365 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-xf9rd_9899b0d7-81a4-49f7-91cd-3c5aa72d49b2/kube-rbac-proxy/0.log" Oct 13 09:38:10 crc kubenswrapper[4685]: I1013 09:38:10.945099 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-xf9rd_9899b0d7-81a4-49f7-91cd-3c5aa72d49b2/manager/0.log" Oct 13 09:38:11 crc kubenswrapper[4685]: I1013 09:38:11.045746 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-f9d897d75-xtnqp_b001d17a-1aea-44ba-86c5-ba6b312156c1/kube-rbac-proxy/0.log" Oct 13 09:38:11 crc kubenswrapper[4685]: I1013 09:38:11.167398 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-f9d897d75-xtnqp_b001d17a-1aea-44ba-86c5-ba6b312156c1/manager/8.log" Oct 13 09:38:11 crc kubenswrapper[4685]: I1013 09:38:11.189170 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-f9d897d75-xtnqp_b001d17a-1aea-44ba-86c5-ba6b312156c1/manager/8.log" Oct 13 09:38:11 crc kubenswrapper[4685]: I1013 09:38:11.394591 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-q8dkt_12dc02ce-76d4-4376-ab6f-b2a75580cc4d/kube-rbac-proxy/0.log" Oct 13 09:38:11 crc kubenswrapper[4685]: I1013 09:38:11.423392 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-q8dkt_12dc02ce-76d4-4376-ab6f-b2a75580cc4d/manager/0.log" Oct 13 09:38:11 crc kubenswrapper[4685]: I1013 09:38:11.581470 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-j65jt_e297e3a5-6cd9-4314-b4dc-76ba4c25c6f5/kube-rbac-proxy/0.log" Oct 13 09:38:11 crc kubenswrapper[4685]: I1013 09:38:11.680836 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-j65jt_e297e3a5-6cd9-4314-b4dc-76ba4c25c6f5/manager/0.log" Oct 13 09:38:11 crc kubenswrapper[4685]: I1013 09:38:11.794698 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-cbwvq_de8c01ad-9e44-42f8-956c-0bb61165a222/kube-rbac-proxy/0.log" Oct 13 09:38:11 crc kubenswrapper[4685]: I1013 09:38:11.874405 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-cbwvq_de8c01ad-9e44-42f8-956c-0bb61165a222/manager/0.log" Oct 13 09:38:11 crc kubenswrapper[4685]: I1013 09:38:11.969033 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-ft77l_d363a932-2a5d-4082-841a-b5d677d0b1a1/kube-rbac-proxy/0.log" Oct 13 09:38:12 crc kubenswrapper[4685]: I1013 09:38:12.154019 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-ft77l_d363a932-2a5d-4082-841a-b5d677d0b1a1/manager/0.log" Oct 13 09:38:12 crc kubenswrapper[4685]: I1013 09:38:12.229351 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-b9jts_3e53276a-ccb3-40a1-b1e6-307cc335ff4d/manager/0.log" Oct 13 09:38:12 crc kubenswrapper[4685]: I1013 09:38:12.263322 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-b9jts_3e53276a-ccb3-40a1-b1e6-307cc335ff4d/kube-rbac-proxy/0.log" Oct 13 09:38:12 crc kubenswrapper[4685]: I1013 09:38:12.415165 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td_d4c8f24c-6e84-4931-8edb-504b184ea7b0/kube-rbac-proxy/0.log" Oct 13 09:38:12 crc kubenswrapper[4685]: I1013 09:38:12.687491 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td_d4c8f24c-6e84-4931-8edb-504b184ea7b0/manager/0.log" Oct 13 09:38:12 crc kubenswrapper[4685]: I1013 09:38:12.885125 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6c8cf6687-nbccj_c31c40db-9982-4692-91cc-26bf7b4ba509/kube-rbac-proxy/0.log" Oct 13 09:38:12 crc kubenswrapper[4685]: I1013 09:38:12.957144 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-859d8f6f74-jwkhb_d2886500-f092-4cba-bd20-50483b96ceb3/kube-rbac-proxy/0.log" Oct 13 09:38:13 crc kubenswrapper[4685]: I1013 09:38:13.196412 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-t7wsk_3d6866fd-e05d-4533-9134-47a83469940b/registry-server/0.log" Oct 13 09:38:13 crc kubenswrapper[4685]: I1013 09:38:13.361761 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6c8cf6687-nbccj_c31c40db-9982-4692-91cc-26bf7b4ba509/operator/0.log" Oct 13 09:38:13 crc kubenswrapper[4685]: I1013 09:38:13.449800 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-7ktz6_a4e11db1-f272-469d-9c22-a649cdbcf95e/kube-rbac-proxy/0.log" Oct 13 09:38:13 crc kubenswrapper[4685]: I1013 09:38:13.588617 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-7ktz6_a4e11db1-f272-469d-9c22-a649cdbcf95e/manager/0.log" Oct 13 09:38:13 crc kubenswrapper[4685]: I1013 09:38:13.712972 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-fhw55_9239cb03-ddca-4542-b188-b89717a00f75/kube-rbac-proxy/0.log" Oct 13 09:38:13 crc kubenswrapper[4685]: I1013 09:38:13.826011 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-fhw55_9239cb03-ddca-4542-b188-b89717a00f75/manager/0.log" Oct 13 09:38:13 crc kubenswrapper[4685]: I1013 09:38:13.938376 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-bgvk8_79c6993d-5a5b-4b38-b678-6c78e639d834/operator/0.log" Oct 13 09:38:14 crc kubenswrapper[4685]: I1013 09:38:14.089626 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-d6gn6_d639bb21-69ae-45c8-8a9c-aac17f57f8dd/kube-rbac-proxy/0.log" Oct 13 09:38:14 crc kubenswrapper[4685]: I1013 09:38:14.146814 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-859d8f6f74-jwkhb_d2886500-f092-4cba-bd20-50483b96ceb3/manager/0.log" Oct 13 09:38:14 crc kubenswrapper[4685]: I1013 09:38:14.228451 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-d6gn6_d639bb21-69ae-45c8-8a9c-aac17f57f8dd/manager/0.log" Oct 13 09:38:14 crc kubenswrapper[4685]: I1013 09:38:14.294259 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-qnmxx_f1085de4-f906-4315-b8b5-a3ee4e7182c1/kube-rbac-proxy/0.log" Oct 13 09:38:14 crc kubenswrapper[4685]: I1013 09:38:14.378358 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-qnmxx_f1085de4-f906-4315-b8b5-a3ee4e7182c1/manager/0.log" Oct 13 09:38:14 crc kubenswrapper[4685]: I1013 09:38:14.444210 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-dd92k_0b4f3fea-3a89-4d28-89c0-436da959e36f/kube-rbac-proxy/0.log" Oct 13 09:38:14 crc kubenswrapper[4685]: I1013 09:38:14.485476 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-dd92k_0b4f3fea-3a89-4d28-89c0-436da959e36f/manager/0.log" Oct 13 09:38:14 crc kubenswrapper[4685]: I1013 09:38:14.563024 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-9wdvd_026b2615-8dc1-4ba7-83d1-1e21f4fa80d2/kube-rbac-proxy/0.log" Oct 13 09:38:14 crc kubenswrapper[4685]: I1013 09:38:14.636484 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-9wdvd_026b2615-8dc1-4ba7-83d1-1e21f4fa80d2/manager/0.log" Oct 13 09:38:15 crc kubenswrapper[4685]: I1013 09:38:15.514555 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:38:15 crc kubenswrapper[4685]: E1013 09:38:15.514887 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:38:19 crc kubenswrapper[4685]: I1013 09:38:19.502686 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:38:19 crc kubenswrapper[4685]: E1013 09:38:19.503323 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:38:26 crc kubenswrapper[4685]: I1013 09:38:26.503747 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:38:26 crc kubenswrapper[4685]: E1013 09:38:26.504568 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:38:30 crc kubenswrapper[4685]: I1013 09:38:30.603420 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-4zvdw_d869fa03-5196-4a23-a7d3-9bb709891678/control-plane-machine-set-operator/0.log" Oct 13 09:38:30 crc kubenswrapper[4685]: I1013 09:38:30.807211 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j6kt6_a185671e-0a3a-4a9e-a884-65b448b1e922/kube-rbac-proxy/0.log" Oct 13 09:38:30 crc kubenswrapper[4685]: I1013 09:38:30.827093 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j6kt6_a185671e-0a3a-4a9e-a884-65b448b1e922/machine-api-operator/0.log" Oct 13 09:38:33 crc kubenswrapper[4685]: I1013 09:38:33.503744 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:38:33 crc kubenswrapper[4685]: E1013 09:38:33.504084 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:38:39 crc kubenswrapper[4685]: I1013 09:38:39.503683 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:38:39 crc kubenswrapper[4685]: E1013 09:38:39.504506 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:38:42 crc kubenswrapper[4685]: I1013 09:38:42.718405 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-k77gp_1fec02c3-da30-45db-9004-73214b0d5a33/cert-manager-controller/0.log" Oct 13 09:38:42 crc kubenswrapper[4685]: I1013 09:38:42.952999 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-8l97z_d1799c36-72e8-4563-b4df-0115cd0e1108/cert-manager-webhook/0.log" Oct 13 09:38:42 crc kubenswrapper[4685]: I1013 09:38:42.972626 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-pw6jw_17313b92-8f5e-45c2-986c-d86cea130b7e/cert-manager-cainjector/0.log" Oct 13 09:38:46 crc kubenswrapper[4685]: I1013 09:38:46.501933 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:38:46 crc kubenswrapper[4685]: E1013 09:38:46.502481 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:38:52 crc kubenswrapper[4685]: I1013 09:38:52.503001 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:38:52 crc kubenswrapper[4685]: E1013 09:38:52.503654 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:38:55 crc kubenswrapper[4685]: I1013 09:38:55.258158 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-zscjx_1c0f8ea9-312a-479f-be32-33028d5d6651/nmstate-console-plugin/0.log" Oct 13 09:38:55 crc kubenswrapper[4685]: I1013 09:38:55.399899 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-5cc6j_ed739611-c4b8-42d8-9c8f-3382d0347c39/nmstate-handler/0.log" Oct 13 09:38:55 crc kubenswrapper[4685]: I1013 09:38:55.449401 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-bgw9w_4cdd7309-1689-4b3a-8165-bda0f92016c9/kube-rbac-proxy/0.log" Oct 13 09:38:55 crc kubenswrapper[4685]: I1013 09:38:55.482515 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-bgw9w_4cdd7309-1689-4b3a-8165-bda0f92016c9/nmstate-metrics/0.log" Oct 13 09:38:55 crc kubenswrapper[4685]: I1013 09:38:55.664331 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-7l6n7_4d407a70-cfdb-427d-94a8-c975db126733/nmstate-operator/0.log" Oct 13 09:38:55 crc kubenswrapper[4685]: I1013 09:38:55.708638 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-66rzd_ac20f0e3-6bff-43fa-93d5-6447bf249314/nmstate-webhook/0.log" Oct 13 09:39:01 crc kubenswrapper[4685]: I1013 09:39:01.503379 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:39:01 crc kubenswrapper[4685]: E1013 09:39:01.504179 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:39:07 crc kubenswrapper[4685]: I1013 09:39:07.503399 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:39:07 crc kubenswrapper[4685]: E1013 09:39:07.504182 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:39:10 crc kubenswrapper[4685]: I1013 09:39:10.213938 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-fmsz9_34bb5061-30c8-49f5-8dca-a411b3f128b2/kube-rbac-proxy/0.log" Oct 13 09:39:10 crc kubenswrapper[4685]: I1013 09:39:10.238170 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-fmsz9_34bb5061-30c8-49f5-8dca-a411b3f128b2/controller/0.log" Oct 13 09:39:10 crc kubenswrapper[4685]: I1013 09:39:10.346949 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-frr-files/0.log" Oct 13 09:39:10 crc kubenswrapper[4685]: I1013 09:39:10.595003 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-frr-files/0.log" Oct 13 09:39:10 crc kubenswrapper[4685]: I1013 09:39:10.622650 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-reloader/0.log" Oct 13 09:39:10 crc kubenswrapper[4685]: I1013 09:39:10.639712 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-metrics/0.log" Oct 13 09:39:10 crc kubenswrapper[4685]: I1013 09:39:10.644963 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-reloader/0.log" Oct 13 09:39:10 crc kubenswrapper[4685]: I1013 09:39:10.821964 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-metrics/0.log" Oct 13 09:39:10 crc kubenswrapper[4685]: I1013 09:39:10.826857 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-frr-files/0.log" Oct 13 09:39:10 crc kubenswrapper[4685]: I1013 09:39:10.859381 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-reloader/0.log" Oct 13 09:39:10 crc kubenswrapper[4685]: I1013 09:39:10.907537 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-metrics/0.log" Oct 13 09:39:11 crc kubenswrapper[4685]: I1013 09:39:11.066072 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-reloader/0.log" Oct 13 09:39:11 crc kubenswrapper[4685]: I1013 09:39:11.118365 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-metrics/0.log" Oct 13 09:39:11 crc kubenswrapper[4685]: I1013 09:39:11.125568 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-frr-files/0.log" Oct 13 09:39:11 crc kubenswrapper[4685]: I1013 09:39:11.138899 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/controller/0.log" Oct 13 09:39:11 crc kubenswrapper[4685]: I1013 09:39:11.300248 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/frr-metrics/0.log" Oct 13 09:39:11 crc kubenswrapper[4685]: I1013 09:39:11.362411 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/kube-rbac-proxy/0.log" Oct 13 09:39:11 crc kubenswrapper[4685]: I1013 09:39:11.485927 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/kube-rbac-proxy-frr/0.log" Oct 13 09:39:11 crc kubenswrapper[4685]: I1013 09:39:11.606545 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/reloader/0.log" Oct 13 09:39:11 crc kubenswrapper[4685]: I1013 09:39:11.800825 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-fkd9z_9b097f6c-ec81-4342-b4c1-63520267ba08/frr-k8s-webhook-server/0.log" Oct 13 09:39:11 crc kubenswrapper[4685]: I1013 09:39:11.996274 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-759fc95d5d-79ckz_b3d2fe67-48db-4dbc-869d-d0c13b18ec8a/manager/0.log" Oct 13 09:39:12 crc kubenswrapper[4685]: I1013 09:39:12.263959 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6f5fffbd86-pw7w6_ce259f59-820d-41e4-8ef7-3b977664b7ea/webhook-server/0.log" Oct 13 09:39:12 crc kubenswrapper[4685]: I1013 09:39:12.320272 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-ddb2s_55ccd123-f671-4230-b2d5-e6ffb265429a/kube-rbac-proxy/0.log" Oct 13 09:39:12 crc kubenswrapper[4685]: I1013 09:39:12.566762 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/frr/0.log" Oct 13 09:39:12 crc kubenswrapper[4685]: I1013 09:39:12.820291 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-ddb2s_55ccd123-f671-4230-b2d5-e6ffb265429a/speaker/0.log" Oct 13 09:39:16 crc kubenswrapper[4685]: I1013 09:39:16.503182 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:39:16 crc kubenswrapper[4685]: E1013 09:39:16.503791 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:39:20 crc kubenswrapper[4685]: I1013 09:39:20.502570 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:39:20 crc kubenswrapper[4685]: E1013 09:39:20.503548 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:39:24 crc kubenswrapper[4685]: I1013 09:39:24.619354 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z_624e992c-aa90-41ff-a746-adf45015b81a/util/0.log" Oct 13 09:39:24 crc kubenswrapper[4685]: I1013 09:39:24.820201 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z_624e992c-aa90-41ff-a746-adf45015b81a/pull/0.log" Oct 13 09:39:24 crc kubenswrapper[4685]: I1013 09:39:24.865211 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z_624e992c-aa90-41ff-a746-adf45015b81a/util/0.log" Oct 13 09:39:24 crc kubenswrapper[4685]: I1013 09:39:24.874881 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z_624e992c-aa90-41ff-a746-adf45015b81a/pull/0.log" Oct 13 09:39:25 crc kubenswrapper[4685]: I1013 09:39:25.067418 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z_624e992c-aa90-41ff-a746-adf45015b81a/util/0.log" Oct 13 09:39:25 crc kubenswrapper[4685]: I1013 09:39:25.104690 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z_624e992c-aa90-41ff-a746-adf45015b81a/pull/0.log" Oct 13 09:39:25 crc kubenswrapper[4685]: I1013 09:39:25.144220 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z_624e992c-aa90-41ff-a746-adf45015b81a/extract/0.log" Oct 13 09:39:25 crc kubenswrapper[4685]: I1013 09:39:25.326965 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-flhsp_4f92f2fa-5809-4485-b5e1-90e6ebf8efd0/extract-utilities/0.log" Oct 13 09:39:25 crc kubenswrapper[4685]: I1013 09:39:25.515436 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-flhsp_4f92f2fa-5809-4485-b5e1-90e6ebf8efd0/extract-utilities/0.log" Oct 13 09:39:25 crc kubenswrapper[4685]: I1013 09:39:25.524697 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-flhsp_4f92f2fa-5809-4485-b5e1-90e6ebf8efd0/extract-content/0.log" Oct 13 09:39:25 crc kubenswrapper[4685]: I1013 09:39:25.557612 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-flhsp_4f92f2fa-5809-4485-b5e1-90e6ebf8efd0/extract-content/0.log" Oct 13 09:39:25 crc kubenswrapper[4685]: I1013 09:39:25.689037 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-flhsp_4f92f2fa-5809-4485-b5e1-90e6ebf8efd0/extract-utilities/0.log" Oct 13 09:39:25 crc kubenswrapper[4685]: I1013 09:39:25.761687 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-flhsp_4f92f2fa-5809-4485-b5e1-90e6ebf8efd0/extract-content/0.log" Oct 13 09:39:26 crc kubenswrapper[4685]: I1013 09:39:26.058565 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9z22p_d4dc80bb-b601-4804-80f0-e3f97493b377/extract-utilities/0.log" Oct 13 09:39:26 crc kubenswrapper[4685]: I1013 09:39:26.211583 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-flhsp_4f92f2fa-5809-4485-b5e1-90e6ebf8efd0/registry-server/0.log" Oct 13 09:39:26 crc kubenswrapper[4685]: I1013 09:39:26.244003 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9z22p_d4dc80bb-b601-4804-80f0-e3f97493b377/extract-content/0.log" Oct 13 09:39:26 crc kubenswrapper[4685]: I1013 09:39:26.277128 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9z22p_d4dc80bb-b601-4804-80f0-e3f97493b377/extract-utilities/0.log" Oct 13 09:39:26 crc kubenswrapper[4685]: I1013 09:39:26.326640 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9z22p_d4dc80bb-b601-4804-80f0-e3f97493b377/extract-content/0.log" Oct 13 09:39:26 crc kubenswrapper[4685]: I1013 09:39:26.542807 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9z22p_d4dc80bb-b601-4804-80f0-e3f97493b377/extract-utilities/0.log" Oct 13 09:39:26 crc kubenswrapper[4685]: I1013 09:39:26.594964 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9z22p_d4dc80bb-b601-4804-80f0-e3f97493b377/extract-content/0.log" Oct 13 09:39:26 crc kubenswrapper[4685]: I1013 09:39:26.791204 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc_207d7e9f-5cfb-4476-962f-8f6b960d32d2/util/0.log" Oct 13 09:39:27 crc kubenswrapper[4685]: I1013 09:39:27.163634 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc_207d7e9f-5cfb-4476-962f-8f6b960d32d2/pull/0.log" Oct 13 09:39:27 crc kubenswrapper[4685]: I1013 09:39:27.224614 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc_207d7e9f-5cfb-4476-962f-8f6b960d32d2/pull/0.log" Oct 13 09:39:27 crc kubenswrapper[4685]: I1013 09:39:27.229423 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9z22p_d4dc80bb-b601-4804-80f0-e3f97493b377/registry-server/0.log" Oct 13 09:39:27 crc kubenswrapper[4685]: I1013 09:39:27.262646 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc_207d7e9f-5cfb-4476-962f-8f6b960d32d2/util/0.log" Oct 13 09:39:27 crc kubenswrapper[4685]: I1013 09:39:27.465004 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc_207d7e9f-5cfb-4476-962f-8f6b960d32d2/pull/0.log" Oct 13 09:39:27 crc kubenswrapper[4685]: I1013 09:39:27.483269 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc_207d7e9f-5cfb-4476-962f-8f6b960d32d2/extract/0.log" Oct 13 09:39:27 crc kubenswrapper[4685]: I1013 09:39:27.486959 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc_207d7e9f-5cfb-4476-962f-8f6b960d32d2/util/0.log" Oct 13 09:39:27 crc kubenswrapper[4685]: I1013 09:39:27.692125 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-5pssn_30028087-7eef-4beb-8011-304f2dce2e01/marketplace-operator/0.log" Oct 13 09:39:27 crc kubenswrapper[4685]: I1013 09:39:27.740211 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pfbfb_396234a0-0c33-45f1-9e9a-a52c6e69e7c0/extract-utilities/0.log" Oct 13 09:39:27 crc kubenswrapper[4685]: I1013 09:39:27.946070 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pfbfb_396234a0-0c33-45f1-9e9a-a52c6e69e7c0/extract-content/0.log" Oct 13 09:39:27 crc kubenswrapper[4685]: I1013 09:39:27.966302 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pfbfb_396234a0-0c33-45f1-9e9a-a52c6e69e7c0/extract-content/0.log" Oct 13 09:39:28 crc kubenswrapper[4685]: I1013 09:39:28.013433 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pfbfb_396234a0-0c33-45f1-9e9a-a52c6e69e7c0/extract-utilities/0.log" Oct 13 09:39:28 crc kubenswrapper[4685]: I1013 09:39:28.221571 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pfbfb_396234a0-0c33-45f1-9e9a-a52c6e69e7c0/extract-content/0.log" Oct 13 09:39:28 crc kubenswrapper[4685]: I1013 09:39:28.290015 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pfbfb_396234a0-0c33-45f1-9e9a-a52c6e69e7c0/extract-utilities/0.log" Oct 13 09:39:28 crc kubenswrapper[4685]: I1013 09:39:28.337242 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pfbfb_396234a0-0c33-45f1-9e9a-a52c6e69e7c0/registry-server/0.log" Oct 13 09:39:28 crc kubenswrapper[4685]: I1013 09:39:28.429480 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t9v8k_9adb7fcf-2d97-4c07-85cf-86cef286dc69/extract-utilities/0.log" Oct 13 09:39:28 crc kubenswrapper[4685]: I1013 09:39:28.603285 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t9v8k_9adb7fcf-2d97-4c07-85cf-86cef286dc69/extract-utilities/0.log" Oct 13 09:39:28 crc kubenswrapper[4685]: I1013 09:39:28.660707 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t9v8k_9adb7fcf-2d97-4c07-85cf-86cef286dc69/extract-content/0.log" Oct 13 09:39:28 crc kubenswrapper[4685]: I1013 09:39:28.675756 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t9v8k_9adb7fcf-2d97-4c07-85cf-86cef286dc69/extract-content/0.log" Oct 13 09:39:28 crc kubenswrapper[4685]: I1013 09:39:28.865125 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t9v8k_9adb7fcf-2d97-4c07-85cf-86cef286dc69/extract-content/0.log" Oct 13 09:39:28 crc kubenswrapper[4685]: I1013 09:39:28.883054 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t9v8k_9adb7fcf-2d97-4c07-85cf-86cef286dc69/extract-utilities/0.log" Oct 13 09:39:29 crc kubenswrapper[4685]: I1013 09:39:29.443932 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t9v8k_9adb7fcf-2d97-4c07-85cf-86cef286dc69/registry-server/0.log" Oct 13 09:39:30 crc kubenswrapper[4685]: I1013 09:39:30.503602 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:39:30 crc kubenswrapper[4685]: E1013 09:39:30.504120 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:39:34 crc kubenswrapper[4685]: I1013 09:39:34.503674 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:39:34 crc kubenswrapper[4685]: E1013 09:39:34.504313 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:39:43 crc kubenswrapper[4685]: I1013 09:39:43.502419 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:39:43 crc kubenswrapper[4685]: E1013 09:39:43.503237 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:39:46 crc kubenswrapper[4685]: I1013 09:39:46.502908 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:39:46 crc kubenswrapper[4685]: E1013 09:39:46.503371 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:39:57 crc kubenswrapper[4685]: I1013 09:39:57.503654 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:39:57 crc kubenswrapper[4685]: E1013 09:39:57.504370 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:40:01 crc kubenswrapper[4685]: I1013 09:40:01.502710 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:40:01 crc kubenswrapper[4685]: E1013 09:40:01.504628 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:40:06 crc kubenswrapper[4685]: E1013 09:40:06.817570 4685 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.65:50602->38.102.83.65:35759: write tcp 38.102.83.65:50602->38.102.83.65:35759: write: broken pipe Oct 13 09:40:12 crc kubenswrapper[4685]: I1013 09:40:12.504205 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:40:12 crc kubenswrapper[4685]: E1013 09:40:12.505082 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:40:12 crc kubenswrapper[4685]: I1013 09:40:12.505281 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:40:12 crc kubenswrapper[4685]: E1013 09:40:12.505517 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:40:24 crc kubenswrapper[4685]: I1013 09:40:24.503521 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:40:24 crc kubenswrapper[4685]: I1013 09:40:24.504327 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:40:24 crc kubenswrapper[4685]: E1013 09:40:24.504481 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:40:24 crc kubenswrapper[4685]: E1013 09:40:24.504792 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:40:27 crc kubenswrapper[4685]: I1013 09:40:27.937279 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cxmcr"] Oct 13 09:40:27 crc kubenswrapper[4685]: E1013 09:40:27.938724 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cadad5a8-4707-484a-9320-7e4f7875b96f" containerName="container-00" Oct 13 09:40:27 crc kubenswrapper[4685]: I1013 09:40:27.938837 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="cadad5a8-4707-484a-9320-7e4f7875b96f" containerName="container-00" Oct 13 09:40:27 crc kubenswrapper[4685]: I1013 09:40:27.939206 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="cadad5a8-4707-484a-9320-7e4f7875b96f" containerName="container-00" Oct 13 09:40:27 crc kubenswrapper[4685]: I1013 09:40:27.940874 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cxmcr" Oct 13 09:40:27 crc kubenswrapper[4685]: I1013 09:40:27.953390 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cxmcr"] Oct 13 09:40:28 crc kubenswrapper[4685]: I1013 09:40:28.095799 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c-utilities\") pod \"certified-operators-cxmcr\" (UID: \"65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c\") " pod="openshift-marketplace/certified-operators-cxmcr" Oct 13 09:40:28 crc kubenswrapper[4685]: I1013 09:40:28.096145 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c-catalog-content\") pod \"certified-operators-cxmcr\" (UID: \"65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c\") " pod="openshift-marketplace/certified-operators-cxmcr" Oct 13 09:40:28 crc kubenswrapper[4685]: I1013 09:40:28.096259 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9l8f\" (UniqueName: \"kubernetes.io/projected/65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c-kube-api-access-n9l8f\") pod \"certified-operators-cxmcr\" (UID: \"65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c\") " pod="openshift-marketplace/certified-operators-cxmcr" Oct 13 09:40:28 crc kubenswrapper[4685]: I1013 09:40:28.198780 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c-utilities\") pod \"certified-operators-cxmcr\" (UID: \"65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c\") " pod="openshift-marketplace/certified-operators-cxmcr" Oct 13 09:40:28 crc kubenswrapper[4685]: I1013 09:40:28.198938 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c-catalog-content\") pod \"certified-operators-cxmcr\" (UID: \"65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c\") " pod="openshift-marketplace/certified-operators-cxmcr" Oct 13 09:40:28 crc kubenswrapper[4685]: I1013 09:40:28.199404 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c-catalog-content\") pod \"certified-operators-cxmcr\" (UID: \"65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c\") " pod="openshift-marketplace/certified-operators-cxmcr" Oct 13 09:40:28 crc kubenswrapper[4685]: I1013 09:40:28.199415 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c-utilities\") pod \"certified-operators-cxmcr\" (UID: \"65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c\") " pod="openshift-marketplace/certified-operators-cxmcr" Oct 13 09:40:28 crc kubenswrapper[4685]: I1013 09:40:28.199469 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9l8f\" (UniqueName: \"kubernetes.io/projected/65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c-kube-api-access-n9l8f\") pod \"certified-operators-cxmcr\" (UID: \"65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c\") " pod="openshift-marketplace/certified-operators-cxmcr" Oct 13 09:40:28 crc kubenswrapper[4685]: I1013 09:40:28.220133 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9l8f\" (UniqueName: \"kubernetes.io/projected/65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c-kube-api-access-n9l8f\") pod \"certified-operators-cxmcr\" (UID: \"65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c\") " pod="openshift-marketplace/certified-operators-cxmcr" Oct 13 09:40:28 crc kubenswrapper[4685]: I1013 09:40:28.284351 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cxmcr" Oct 13 09:40:28 crc kubenswrapper[4685]: I1013 09:40:28.913239 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cxmcr"] Oct 13 09:40:29 crc kubenswrapper[4685]: I1013 09:40:29.303219 4685 generic.go:334] "Generic (PLEG): container finished" podID="65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c" containerID="7844bd413657575e6e76febf02d3ae451c24daf888ebd718aee039d156c7cdbe" exitCode=0 Oct 13 09:40:29 crc kubenswrapper[4685]: I1013 09:40:29.303319 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cxmcr" event={"ID":"65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c","Type":"ContainerDied","Data":"7844bd413657575e6e76febf02d3ae451c24daf888ebd718aee039d156c7cdbe"} Oct 13 09:40:29 crc kubenswrapper[4685]: I1013 09:40:29.303509 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cxmcr" event={"ID":"65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c","Type":"ContainerStarted","Data":"273d3a6fdb3d94cdfa9b898ab45bcfdf987ff7eca8eb1004df74f8fdc1664e27"} Oct 13 09:40:29 crc kubenswrapper[4685]: I1013 09:40:29.305507 4685 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 13 09:40:31 crc kubenswrapper[4685]: I1013 09:40:31.321539 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cxmcr" event={"ID":"65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c","Type":"ContainerStarted","Data":"ee32655d49a8134905b12f1bbfc75205ea9c92273d41602ae383d5efc376b0ab"} Oct 13 09:40:32 crc kubenswrapper[4685]: I1013 09:40:32.339857 4685 generic.go:334] "Generic (PLEG): container finished" podID="65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c" containerID="ee32655d49a8134905b12f1bbfc75205ea9c92273d41602ae383d5efc376b0ab" exitCode=0 Oct 13 09:40:32 crc kubenswrapper[4685]: I1013 09:40:32.339952 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cxmcr" event={"ID":"65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c","Type":"ContainerDied","Data":"ee32655d49a8134905b12f1bbfc75205ea9c92273d41602ae383d5efc376b0ab"} Oct 13 09:40:33 crc kubenswrapper[4685]: I1013 09:40:33.351765 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cxmcr" event={"ID":"65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c","Type":"ContainerStarted","Data":"d12104003d465dd133087ce839318940ba72e1a61e1c3ccc49eccf3b8be365fe"} Oct 13 09:40:33 crc kubenswrapper[4685]: I1013 09:40:33.375011 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cxmcr" podStartSLOduration=2.868655019 podStartE2EDuration="6.374991997s" podCreationTimestamp="2025-10-13 09:40:27 +0000 UTC" firstStartedPulling="2025-10-13 09:40:29.305264491 +0000 UTC m=+3354.453140252" lastFinishedPulling="2025-10-13 09:40:32.811601469 +0000 UTC m=+3357.959477230" observedRunningTime="2025-10-13 09:40:33.370481625 +0000 UTC m=+3358.518357396" watchObservedRunningTime="2025-10-13 09:40:33.374991997 +0000 UTC m=+3358.522867778" Oct 13 09:40:35 crc kubenswrapper[4685]: I1013 09:40:35.512534 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:40:35 crc kubenswrapper[4685]: E1013 09:40:35.514409 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:40:37 crc kubenswrapper[4685]: I1013 09:40:37.502544 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:40:37 crc kubenswrapper[4685]: E1013 09:40:37.503121 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:40:38 crc kubenswrapper[4685]: I1013 09:40:38.286211 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cxmcr" Oct 13 09:40:38 crc kubenswrapper[4685]: I1013 09:40:38.286270 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cxmcr" Oct 13 09:40:38 crc kubenswrapper[4685]: I1013 09:40:38.360222 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cxmcr" Oct 13 09:40:38 crc kubenswrapper[4685]: I1013 09:40:38.465796 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cxmcr" Oct 13 09:40:38 crc kubenswrapper[4685]: I1013 09:40:38.600994 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cxmcr"] Oct 13 09:40:40 crc kubenswrapper[4685]: I1013 09:40:40.417668 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cxmcr" podUID="65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c" containerName="registry-server" containerID="cri-o://d12104003d465dd133087ce839318940ba72e1a61e1c3ccc49eccf3b8be365fe" gracePeriod=2 Oct 13 09:40:40 crc kubenswrapper[4685]: I1013 09:40:40.901851 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cxmcr" Oct 13 09:40:40 crc kubenswrapper[4685]: I1013 09:40:40.952963 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c-utilities\") pod \"65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c\" (UID: \"65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c\") " Oct 13 09:40:40 crc kubenswrapper[4685]: I1013 09:40:40.953179 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9l8f\" (UniqueName: \"kubernetes.io/projected/65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c-kube-api-access-n9l8f\") pod \"65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c\" (UID: \"65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c\") " Oct 13 09:40:40 crc kubenswrapper[4685]: I1013 09:40:40.953223 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c-catalog-content\") pod \"65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c\" (UID: \"65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c\") " Oct 13 09:40:40 crc kubenswrapper[4685]: I1013 09:40:40.953909 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c-utilities" (OuterVolumeSpecName: "utilities") pod "65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c" (UID: "65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:40:40 crc kubenswrapper[4685]: I1013 09:40:40.959950 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c-kube-api-access-n9l8f" (OuterVolumeSpecName: "kube-api-access-n9l8f") pod "65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c" (UID: "65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c"). InnerVolumeSpecName "kube-api-access-n9l8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:40:41 crc kubenswrapper[4685]: I1013 09:40:41.006808 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c" (UID: "65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:40:41 crc kubenswrapper[4685]: I1013 09:40:41.055926 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 09:40:41 crc kubenswrapper[4685]: I1013 09:40:41.055960 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9l8f\" (UniqueName: \"kubernetes.io/projected/65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c-kube-api-access-n9l8f\") on node \"crc\" DevicePath \"\"" Oct 13 09:40:41 crc kubenswrapper[4685]: I1013 09:40:41.055970 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 09:40:41 crc kubenswrapper[4685]: I1013 09:40:41.428709 4685 generic.go:334] "Generic (PLEG): container finished" podID="65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c" containerID="d12104003d465dd133087ce839318940ba72e1a61e1c3ccc49eccf3b8be365fe" exitCode=0 Oct 13 09:40:41 crc kubenswrapper[4685]: I1013 09:40:41.428767 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cxmcr" Oct 13 09:40:41 crc kubenswrapper[4685]: I1013 09:40:41.428797 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cxmcr" event={"ID":"65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c","Type":"ContainerDied","Data":"d12104003d465dd133087ce839318940ba72e1a61e1c3ccc49eccf3b8be365fe"} Oct 13 09:40:41 crc kubenswrapper[4685]: I1013 09:40:41.430005 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cxmcr" event={"ID":"65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c","Type":"ContainerDied","Data":"273d3a6fdb3d94cdfa9b898ab45bcfdf987ff7eca8eb1004df74f8fdc1664e27"} Oct 13 09:40:41 crc kubenswrapper[4685]: I1013 09:40:41.430040 4685 scope.go:117] "RemoveContainer" containerID="d12104003d465dd133087ce839318940ba72e1a61e1c3ccc49eccf3b8be365fe" Oct 13 09:40:41 crc kubenswrapper[4685]: I1013 09:40:41.458868 4685 scope.go:117] "RemoveContainer" containerID="ee32655d49a8134905b12f1bbfc75205ea9c92273d41602ae383d5efc376b0ab" Oct 13 09:40:41 crc kubenswrapper[4685]: I1013 09:40:41.480485 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cxmcr"] Oct 13 09:40:41 crc kubenswrapper[4685]: I1013 09:40:41.497678 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cxmcr"] Oct 13 09:40:41 crc kubenswrapper[4685]: I1013 09:40:41.500248 4685 scope.go:117] "RemoveContainer" containerID="7844bd413657575e6e76febf02d3ae451c24daf888ebd718aee039d156c7cdbe" Oct 13 09:40:41 crc kubenswrapper[4685]: I1013 09:40:41.527432 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c" path="/var/lib/kubelet/pods/65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c/volumes" Oct 13 09:40:41 crc kubenswrapper[4685]: I1013 09:40:41.598242 4685 scope.go:117] "RemoveContainer" containerID="d12104003d465dd133087ce839318940ba72e1a61e1c3ccc49eccf3b8be365fe" Oct 13 09:40:41 crc kubenswrapper[4685]: E1013 09:40:41.598703 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d12104003d465dd133087ce839318940ba72e1a61e1c3ccc49eccf3b8be365fe\": container with ID starting with d12104003d465dd133087ce839318940ba72e1a61e1c3ccc49eccf3b8be365fe not found: ID does not exist" containerID="d12104003d465dd133087ce839318940ba72e1a61e1c3ccc49eccf3b8be365fe" Oct 13 09:40:41 crc kubenswrapper[4685]: I1013 09:40:41.598730 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d12104003d465dd133087ce839318940ba72e1a61e1c3ccc49eccf3b8be365fe"} err="failed to get container status \"d12104003d465dd133087ce839318940ba72e1a61e1c3ccc49eccf3b8be365fe\": rpc error: code = NotFound desc = could not find container \"d12104003d465dd133087ce839318940ba72e1a61e1c3ccc49eccf3b8be365fe\": container with ID starting with d12104003d465dd133087ce839318940ba72e1a61e1c3ccc49eccf3b8be365fe not found: ID does not exist" Oct 13 09:40:41 crc kubenswrapper[4685]: I1013 09:40:41.598748 4685 scope.go:117] "RemoveContainer" containerID="ee32655d49a8134905b12f1bbfc75205ea9c92273d41602ae383d5efc376b0ab" Oct 13 09:40:41 crc kubenswrapper[4685]: E1013 09:40:41.599048 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee32655d49a8134905b12f1bbfc75205ea9c92273d41602ae383d5efc376b0ab\": container with ID starting with ee32655d49a8134905b12f1bbfc75205ea9c92273d41602ae383d5efc376b0ab not found: ID does not exist" containerID="ee32655d49a8134905b12f1bbfc75205ea9c92273d41602ae383d5efc376b0ab" Oct 13 09:40:41 crc kubenswrapper[4685]: I1013 09:40:41.599066 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee32655d49a8134905b12f1bbfc75205ea9c92273d41602ae383d5efc376b0ab"} err="failed to get container status \"ee32655d49a8134905b12f1bbfc75205ea9c92273d41602ae383d5efc376b0ab\": rpc error: code = NotFound desc = could not find container \"ee32655d49a8134905b12f1bbfc75205ea9c92273d41602ae383d5efc376b0ab\": container with ID starting with ee32655d49a8134905b12f1bbfc75205ea9c92273d41602ae383d5efc376b0ab not found: ID does not exist" Oct 13 09:40:41 crc kubenswrapper[4685]: I1013 09:40:41.599078 4685 scope.go:117] "RemoveContainer" containerID="7844bd413657575e6e76febf02d3ae451c24daf888ebd718aee039d156c7cdbe" Oct 13 09:40:41 crc kubenswrapper[4685]: E1013 09:40:41.599398 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7844bd413657575e6e76febf02d3ae451c24daf888ebd718aee039d156c7cdbe\": container with ID starting with 7844bd413657575e6e76febf02d3ae451c24daf888ebd718aee039d156c7cdbe not found: ID does not exist" containerID="7844bd413657575e6e76febf02d3ae451c24daf888ebd718aee039d156c7cdbe" Oct 13 09:40:41 crc kubenswrapper[4685]: I1013 09:40:41.599417 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7844bd413657575e6e76febf02d3ae451c24daf888ebd718aee039d156c7cdbe"} err="failed to get container status \"7844bd413657575e6e76febf02d3ae451c24daf888ebd718aee039d156c7cdbe\": rpc error: code = NotFound desc = could not find container \"7844bd413657575e6e76febf02d3ae451c24daf888ebd718aee039d156c7cdbe\": container with ID starting with 7844bd413657575e6e76febf02d3ae451c24daf888ebd718aee039d156c7cdbe not found: ID does not exist" Oct 13 09:40:50 crc kubenswrapper[4685]: I1013 09:40:50.503442 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:40:50 crc kubenswrapper[4685]: E1013 09:40:50.504327 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:40:51 crc kubenswrapper[4685]: I1013 09:40:51.502872 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:40:51 crc kubenswrapper[4685]: E1013 09:40:51.503487 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:41:03 crc kubenswrapper[4685]: I1013 09:41:03.503177 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:41:03 crc kubenswrapper[4685]: E1013 09:41:03.504037 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:41:06 crc kubenswrapper[4685]: I1013 09:41:06.503757 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:41:06 crc kubenswrapper[4685]: E1013 09:41:06.504442 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:41:16 crc kubenswrapper[4685]: I1013 09:41:16.502581 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:41:16 crc kubenswrapper[4685]: E1013 09:41:16.503637 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:41:17 crc kubenswrapper[4685]: I1013 09:41:17.503241 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:41:17 crc kubenswrapper[4685]: E1013 09:41:17.504452 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:41:21 crc kubenswrapper[4685]: I1013 09:41:21.934461 4685 generic.go:334] "Generic (PLEG): container finished" podID="60144059-a904-431c-bf75-12f923f281f8" containerID="e456529df4d346d29295c8c35f85e2d7edb3c8ebf1e431f09349bccd91d5996b" exitCode=0 Oct 13 09:41:21 crc kubenswrapper[4685]: I1013 09:41:21.934580 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cv7wl/must-gather-vjvgz" event={"ID":"60144059-a904-431c-bf75-12f923f281f8","Type":"ContainerDied","Data":"e456529df4d346d29295c8c35f85e2d7edb3c8ebf1e431f09349bccd91d5996b"} Oct 13 09:41:21 crc kubenswrapper[4685]: I1013 09:41:21.936626 4685 scope.go:117] "RemoveContainer" containerID="e456529df4d346d29295c8c35f85e2d7edb3c8ebf1e431f09349bccd91d5996b" Oct 13 09:41:22 crc kubenswrapper[4685]: I1013 09:41:22.407069 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-cv7wl_must-gather-vjvgz_60144059-a904-431c-bf75-12f923f281f8/gather/0.log" Oct 13 09:41:27 crc kubenswrapper[4685]: I1013 09:41:27.503629 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:41:27 crc kubenswrapper[4685]: E1013 09:41:27.504556 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:41:31 crc kubenswrapper[4685]: I1013 09:41:31.333429 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-cv7wl/must-gather-vjvgz"] Oct 13 09:41:31 crc kubenswrapper[4685]: I1013 09:41:31.334062 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-cv7wl/must-gather-vjvgz" podUID="60144059-a904-431c-bf75-12f923f281f8" containerName="copy" containerID="cri-o://a839318753cc0e31e5b657f58bd63b2f3948800873127f682fb52b95d77e2eb9" gracePeriod=2 Oct 13 09:41:31 crc kubenswrapper[4685]: I1013 09:41:31.350990 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-cv7wl/must-gather-vjvgz"] Oct 13 09:41:31 crc kubenswrapper[4685]: I1013 09:41:31.502993 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:41:31 crc kubenswrapper[4685]: E1013 09:41:31.503568 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:41:31 crc kubenswrapper[4685]: I1013 09:41:31.777069 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-cv7wl_must-gather-vjvgz_60144059-a904-431c-bf75-12f923f281f8/copy/0.log" Oct 13 09:41:31 crc kubenswrapper[4685]: I1013 09:41:31.777578 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cv7wl/must-gather-vjvgz" Oct 13 09:41:31 crc kubenswrapper[4685]: I1013 09:41:31.815661 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bn5fp\" (UniqueName: \"kubernetes.io/projected/60144059-a904-431c-bf75-12f923f281f8-kube-api-access-bn5fp\") pod \"60144059-a904-431c-bf75-12f923f281f8\" (UID: \"60144059-a904-431c-bf75-12f923f281f8\") " Oct 13 09:41:31 crc kubenswrapper[4685]: I1013 09:41:31.815821 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/60144059-a904-431c-bf75-12f923f281f8-must-gather-output\") pod \"60144059-a904-431c-bf75-12f923f281f8\" (UID: \"60144059-a904-431c-bf75-12f923f281f8\") " Oct 13 09:41:31 crc kubenswrapper[4685]: I1013 09:41:31.820575 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60144059-a904-431c-bf75-12f923f281f8-kube-api-access-bn5fp" (OuterVolumeSpecName: "kube-api-access-bn5fp") pod "60144059-a904-431c-bf75-12f923f281f8" (UID: "60144059-a904-431c-bf75-12f923f281f8"). InnerVolumeSpecName "kube-api-access-bn5fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:41:31 crc kubenswrapper[4685]: I1013 09:41:31.917836 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bn5fp\" (UniqueName: \"kubernetes.io/projected/60144059-a904-431c-bf75-12f923f281f8-kube-api-access-bn5fp\") on node \"crc\" DevicePath \"\"" Oct 13 09:41:31 crc kubenswrapper[4685]: I1013 09:41:31.955719 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60144059-a904-431c-bf75-12f923f281f8-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "60144059-a904-431c-bf75-12f923f281f8" (UID: "60144059-a904-431c-bf75-12f923f281f8"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:41:32 crc kubenswrapper[4685]: I1013 09:41:32.019881 4685 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/60144059-a904-431c-bf75-12f923f281f8-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 13 09:41:32 crc kubenswrapper[4685]: I1013 09:41:32.026374 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-cv7wl_must-gather-vjvgz_60144059-a904-431c-bf75-12f923f281f8/copy/0.log" Oct 13 09:41:32 crc kubenswrapper[4685]: I1013 09:41:32.026676 4685 generic.go:334] "Generic (PLEG): container finished" podID="60144059-a904-431c-bf75-12f923f281f8" containerID="a839318753cc0e31e5b657f58bd63b2f3948800873127f682fb52b95d77e2eb9" exitCode=143 Oct 13 09:41:32 crc kubenswrapper[4685]: I1013 09:41:32.026721 4685 scope.go:117] "RemoveContainer" containerID="a839318753cc0e31e5b657f58bd63b2f3948800873127f682fb52b95d77e2eb9" Oct 13 09:41:32 crc kubenswrapper[4685]: I1013 09:41:32.026823 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cv7wl/must-gather-vjvgz" Oct 13 09:41:32 crc kubenswrapper[4685]: I1013 09:41:32.057062 4685 scope.go:117] "RemoveContainer" containerID="e456529df4d346d29295c8c35f85e2d7edb3c8ebf1e431f09349bccd91d5996b" Oct 13 09:41:32 crc kubenswrapper[4685]: I1013 09:41:32.121470 4685 scope.go:117] "RemoveContainer" containerID="a839318753cc0e31e5b657f58bd63b2f3948800873127f682fb52b95d77e2eb9" Oct 13 09:41:32 crc kubenswrapper[4685]: E1013 09:41:32.125320 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a839318753cc0e31e5b657f58bd63b2f3948800873127f682fb52b95d77e2eb9\": container with ID starting with a839318753cc0e31e5b657f58bd63b2f3948800873127f682fb52b95d77e2eb9 not found: ID does not exist" containerID="a839318753cc0e31e5b657f58bd63b2f3948800873127f682fb52b95d77e2eb9" Oct 13 09:41:32 crc kubenswrapper[4685]: I1013 09:41:32.125364 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a839318753cc0e31e5b657f58bd63b2f3948800873127f682fb52b95d77e2eb9"} err="failed to get container status \"a839318753cc0e31e5b657f58bd63b2f3948800873127f682fb52b95d77e2eb9\": rpc error: code = NotFound desc = could not find container \"a839318753cc0e31e5b657f58bd63b2f3948800873127f682fb52b95d77e2eb9\": container with ID starting with a839318753cc0e31e5b657f58bd63b2f3948800873127f682fb52b95d77e2eb9 not found: ID does not exist" Oct 13 09:41:32 crc kubenswrapper[4685]: I1013 09:41:32.125392 4685 scope.go:117] "RemoveContainer" containerID="e456529df4d346d29295c8c35f85e2d7edb3c8ebf1e431f09349bccd91d5996b" Oct 13 09:41:32 crc kubenswrapper[4685]: E1013 09:41:32.125851 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e456529df4d346d29295c8c35f85e2d7edb3c8ebf1e431f09349bccd91d5996b\": container with ID starting with e456529df4d346d29295c8c35f85e2d7edb3c8ebf1e431f09349bccd91d5996b not found: ID does not exist" containerID="e456529df4d346d29295c8c35f85e2d7edb3c8ebf1e431f09349bccd91d5996b" Oct 13 09:41:32 crc kubenswrapper[4685]: I1013 09:41:32.126000 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e456529df4d346d29295c8c35f85e2d7edb3c8ebf1e431f09349bccd91d5996b"} err="failed to get container status \"e456529df4d346d29295c8c35f85e2d7edb3c8ebf1e431f09349bccd91d5996b\": rpc error: code = NotFound desc = could not find container \"e456529df4d346d29295c8c35f85e2d7edb3c8ebf1e431f09349bccd91d5996b\": container with ID starting with e456529df4d346d29295c8c35f85e2d7edb3c8ebf1e431f09349bccd91d5996b not found: ID does not exist" Oct 13 09:41:33 crc kubenswrapper[4685]: I1013 09:41:33.519003 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60144059-a904-431c-bf75-12f923f281f8" path="/var/lib/kubelet/pods/60144059-a904-431c-bf75-12f923f281f8/volumes" Oct 13 09:41:38 crc kubenswrapper[4685]: I1013 09:41:38.504090 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:41:38 crc kubenswrapper[4685]: E1013 09:41:38.505112 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:41:43 crc kubenswrapper[4685]: I1013 09:41:43.503190 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:41:43 crc kubenswrapper[4685]: E1013 09:41:43.503829 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:41:52 crc kubenswrapper[4685]: I1013 09:41:52.504282 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:41:52 crc kubenswrapper[4685]: E1013 09:41:52.505009 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:41:58 crc kubenswrapper[4685]: I1013 09:41:58.503003 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:41:59 crc kubenswrapper[4685]: I1013 09:41:59.332596 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerStarted","Data":"a75682cfcb35d15ac8d83500d0d93801124ebd38b4b8e7001d309f70d4ce10ad"} Oct 13 09:42:06 crc kubenswrapper[4685]: I1013 09:42:06.503162 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:42:06 crc kubenswrapper[4685]: E1013 09:42:06.503890 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:42:18 crc kubenswrapper[4685]: I1013 09:42:18.414180 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mzp24/must-gather-d64pf"] Oct 13 09:42:18 crc kubenswrapper[4685]: E1013 09:42:18.415090 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c" containerName="extract-utilities" Oct 13 09:42:18 crc kubenswrapper[4685]: I1013 09:42:18.415103 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c" containerName="extract-utilities" Oct 13 09:42:18 crc kubenswrapper[4685]: E1013 09:42:18.415117 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c" containerName="registry-server" Oct 13 09:42:18 crc kubenswrapper[4685]: I1013 09:42:18.415122 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c" containerName="registry-server" Oct 13 09:42:18 crc kubenswrapper[4685]: E1013 09:42:18.415138 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60144059-a904-431c-bf75-12f923f281f8" containerName="copy" Oct 13 09:42:18 crc kubenswrapper[4685]: I1013 09:42:18.415145 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="60144059-a904-431c-bf75-12f923f281f8" containerName="copy" Oct 13 09:42:18 crc kubenswrapper[4685]: E1013 09:42:18.415159 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c" containerName="extract-content" Oct 13 09:42:18 crc kubenswrapper[4685]: I1013 09:42:18.415165 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c" containerName="extract-content" Oct 13 09:42:18 crc kubenswrapper[4685]: E1013 09:42:18.415193 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60144059-a904-431c-bf75-12f923f281f8" containerName="gather" Oct 13 09:42:18 crc kubenswrapper[4685]: I1013 09:42:18.415198 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="60144059-a904-431c-bf75-12f923f281f8" containerName="gather" Oct 13 09:42:18 crc kubenswrapper[4685]: I1013 09:42:18.415381 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="60144059-a904-431c-bf75-12f923f281f8" containerName="gather" Oct 13 09:42:18 crc kubenswrapper[4685]: I1013 09:42:18.415397 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="60144059-a904-431c-bf75-12f923f281f8" containerName="copy" Oct 13 09:42:18 crc kubenswrapper[4685]: I1013 09:42:18.415407 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="65b0f3e1-e65f-42c3-8fc3-958cfb23bc6c" containerName="registry-server" Oct 13 09:42:18 crc kubenswrapper[4685]: I1013 09:42:18.416319 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mzp24/must-gather-d64pf" Oct 13 09:42:18 crc kubenswrapper[4685]: I1013 09:42:18.417859 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mzp24"/"default-dockercfg-mnl4w" Oct 13 09:42:18 crc kubenswrapper[4685]: I1013 09:42:18.418542 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mzp24"/"kube-root-ca.crt" Oct 13 09:42:18 crc kubenswrapper[4685]: I1013 09:42:18.418767 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mzp24"/"openshift-service-ca.crt" Oct 13 09:42:18 crc kubenswrapper[4685]: I1013 09:42:18.431367 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mzp24/must-gather-d64pf"] Oct 13 09:42:18 crc kubenswrapper[4685]: I1013 09:42:18.432320 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5fc0e95d-4f82-4d0f-a745-001486489f27-must-gather-output\") pod \"must-gather-d64pf\" (UID: \"5fc0e95d-4f82-4d0f-a745-001486489f27\") " pod="openshift-must-gather-mzp24/must-gather-d64pf" Oct 13 09:42:18 crc kubenswrapper[4685]: I1013 09:42:18.432584 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lbcz\" (UniqueName: \"kubernetes.io/projected/5fc0e95d-4f82-4d0f-a745-001486489f27-kube-api-access-9lbcz\") pod \"must-gather-d64pf\" (UID: \"5fc0e95d-4f82-4d0f-a745-001486489f27\") " pod="openshift-must-gather-mzp24/must-gather-d64pf" Oct 13 09:42:18 crc kubenswrapper[4685]: I1013 09:42:18.534217 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5fc0e95d-4f82-4d0f-a745-001486489f27-must-gather-output\") pod \"must-gather-d64pf\" (UID: \"5fc0e95d-4f82-4d0f-a745-001486489f27\") " pod="openshift-must-gather-mzp24/must-gather-d64pf" Oct 13 09:42:18 crc kubenswrapper[4685]: I1013 09:42:18.534301 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lbcz\" (UniqueName: \"kubernetes.io/projected/5fc0e95d-4f82-4d0f-a745-001486489f27-kube-api-access-9lbcz\") pod \"must-gather-d64pf\" (UID: \"5fc0e95d-4f82-4d0f-a745-001486489f27\") " pod="openshift-must-gather-mzp24/must-gather-d64pf" Oct 13 09:42:18 crc kubenswrapper[4685]: I1013 09:42:18.535064 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5fc0e95d-4f82-4d0f-a745-001486489f27-must-gather-output\") pod \"must-gather-d64pf\" (UID: \"5fc0e95d-4f82-4d0f-a745-001486489f27\") " pod="openshift-must-gather-mzp24/must-gather-d64pf" Oct 13 09:42:18 crc kubenswrapper[4685]: I1013 09:42:18.556188 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lbcz\" (UniqueName: \"kubernetes.io/projected/5fc0e95d-4f82-4d0f-a745-001486489f27-kube-api-access-9lbcz\") pod \"must-gather-d64pf\" (UID: \"5fc0e95d-4f82-4d0f-a745-001486489f27\") " pod="openshift-must-gather-mzp24/must-gather-d64pf" Oct 13 09:42:18 crc kubenswrapper[4685]: I1013 09:42:18.737511 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mzp24/must-gather-d64pf" Oct 13 09:42:19 crc kubenswrapper[4685]: W1013 09:42:19.267852 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5fc0e95d_4f82_4d0f_a745_001486489f27.slice/crio-2eccf9ddad535476c94098948b5dd3d25973f0ab23476fa94633e645600717dc WatchSource:0}: Error finding container 2eccf9ddad535476c94098948b5dd3d25973f0ab23476fa94633e645600717dc: Status 404 returned error can't find the container with id 2eccf9ddad535476c94098948b5dd3d25973f0ab23476fa94633e645600717dc Oct 13 09:42:19 crc kubenswrapper[4685]: I1013 09:42:19.268152 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mzp24/must-gather-d64pf"] Oct 13 09:42:19 crc kubenswrapper[4685]: I1013 09:42:19.502759 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:42:19 crc kubenswrapper[4685]: E1013 09:42:19.503281 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:42:19 crc kubenswrapper[4685]: I1013 09:42:19.559221 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mzp24/must-gather-d64pf" event={"ID":"5fc0e95d-4f82-4d0f-a745-001486489f27","Type":"ContainerStarted","Data":"de9ab10066e4fb70062591e9cee78610af292f02240bc5a8f175ce5d92ab3aa3"} Oct 13 09:42:19 crc kubenswrapper[4685]: I1013 09:42:19.559257 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mzp24/must-gather-d64pf" event={"ID":"5fc0e95d-4f82-4d0f-a745-001486489f27","Type":"ContainerStarted","Data":"2eccf9ddad535476c94098948b5dd3d25973f0ab23476fa94633e645600717dc"} Oct 13 09:42:20 crc kubenswrapper[4685]: I1013 09:42:20.567691 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mzp24/must-gather-d64pf" event={"ID":"5fc0e95d-4f82-4d0f-a745-001486489f27","Type":"ContainerStarted","Data":"cf3fea82d1a3b59f247721e3b5269489c7096919053ec77f06e90300ab2f13f4"} Oct 13 09:42:20 crc kubenswrapper[4685]: I1013 09:42:20.602594 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mzp24/must-gather-d64pf" podStartSLOduration=2.6025773169999997 podStartE2EDuration="2.602577317s" podCreationTimestamp="2025-10-13 09:42:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:42:20.595658159 +0000 UTC m=+3465.743533920" watchObservedRunningTime="2025-10-13 09:42:20.602577317 +0000 UTC m=+3465.750453068" Oct 13 09:42:23 crc kubenswrapper[4685]: I1013 09:42:23.451501 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mzp24/crc-debug-zq72k"] Oct 13 09:42:23 crc kubenswrapper[4685]: I1013 09:42:23.454295 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mzp24/crc-debug-zq72k" Oct 13 09:42:23 crc kubenswrapper[4685]: I1013 09:42:23.635743 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw5ml\" (UniqueName: \"kubernetes.io/projected/fe57c984-5668-4cdf-8faa-1c9057d82f32-kube-api-access-fw5ml\") pod \"crc-debug-zq72k\" (UID: \"fe57c984-5668-4cdf-8faa-1c9057d82f32\") " pod="openshift-must-gather-mzp24/crc-debug-zq72k" Oct 13 09:42:23 crc kubenswrapper[4685]: I1013 09:42:23.635965 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fe57c984-5668-4cdf-8faa-1c9057d82f32-host\") pod \"crc-debug-zq72k\" (UID: \"fe57c984-5668-4cdf-8faa-1c9057d82f32\") " pod="openshift-must-gather-mzp24/crc-debug-zq72k" Oct 13 09:42:23 crc kubenswrapper[4685]: I1013 09:42:23.737578 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fe57c984-5668-4cdf-8faa-1c9057d82f32-host\") pod \"crc-debug-zq72k\" (UID: \"fe57c984-5668-4cdf-8faa-1c9057d82f32\") " pod="openshift-must-gather-mzp24/crc-debug-zq72k" Oct 13 09:42:23 crc kubenswrapper[4685]: I1013 09:42:23.737683 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fw5ml\" (UniqueName: \"kubernetes.io/projected/fe57c984-5668-4cdf-8faa-1c9057d82f32-kube-api-access-fw5ml\") pod \"crc-debug-zq72k\" (UID: \"fe57c984-5668-4cdf-8faa-1c9057d82f32\") " pod="openshift-must-gather-mzp24/crc-debug-zq72k" Oct 13 09:42:23 crc kubenswrapper[4685]: I1013 09:42:23.737692 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fe57c984-5668-4cdf-8faa-1c9057d82f32-host\") pod \"crc-debug-zq72k\" (UID: \"fe57c984-5668-4cdf-8faa-1c9057d82f32\") " pod="openshift-must-gather-mzp24/crc-debug-zq72k" Oct 13 09:42:23 crc kubenswrapper[4685]: I1013 09:42:23.755766 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fw5ml\" (UniqueName: \"kubernetes.io/projected/fe57c984-5668-4cdf-8faa-1c9057d82f32-kube-api-access-fw5ml\") pod \"crc-debug-zq72k\" (UID: \"fe57c984-5668-4cdf-8faa-1c9057d82f32\") " pod="openshift-must-gather-mzp24/crc-debug-zq72k" Oct 13 09:42:23 crc kubenswrapper[4685]: I1013 09:42:23.775856 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mzp24/crc-debug-zq72k" Oct 13 09:42:23 crc kubenswrapper[4685]: W1013 09:42:23.805107 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe57c984_5668_4cdf_8faa_1c9057d82f32.slice/crio-4859b049deefc563623f22fa8f8062f40d13ac7362bfd586f2bcdeeb43249f4a WatchSource:0}: Error finding container 4859b049deefc563623f22fa8f8062f40d13ac7362bfd586f2bcdeeb43249f4a: Status 404 returned error can't find the container with id 4859b049deefc563623f22fa8f8062f40d13ac7362bfd586f2bcdeeb43249f4a Oct 13 09:42:24 crc kubenswrapper[4685]: I1013 09:42:24.620499 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mzp24/crc-debug-zq72k" event={"ID":"fe57c984-5668-4cdf-8faa-1c9057d82f32","Type":"ContainerStarted","Data":"ff08ea7e1627cab5d337438a0e8b47fb01c4bf5f99477d994d07febbd9106ed6"} Oct 13 09:42:24 crc kubenswrapper[4685]: I1013 09:42:24.621039 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mzp24/crc-debug-zq72k" event={"ID":"fe57c984-5668-4cdf-8faa-1c9057d82f32","Type":"ContainerStarted","Data":"4859b049deefc563623f22fa8f8062f40d13ac7362bfd586f2bcdeeb43249f4a"} Oct 13 09:42:24 crc kubenswrapper[4685]: I1013 09:42:24.639245 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mzp24/crc-debug-zq72k" podStartSLOduration=1.6392291810000001 podStartE2EDuration="1.639229181s" podCreationTimestamp="2025-10-13 09:42:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 09:42:24.639130878 +0000 UTC m=+3469.787006639" watchObservedRunningTime="2025-10-13 09:42:24.639229181 +0000 UTC m=+3469.787104942" Oct 13 09:42:32 crc kubenswrapper[4685]: I1013 09:42:32.503413 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:42:32 crc kubenswrapper[4685]: E1013 09:42:32.505507 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:42:47 crc kubenswrapper[4685]: I1013 09:42:47.502817 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:42:47 crc kubenswrapper[4685]: E1013 09:42:47.503641 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:42:57 crc kubenswrapper[4685]: I1013 09:42:57.045773 4685 scope.go:117] "RemoveContainer" containerID="5b7df5bcacf6cf5af9ee108468642b8914d6b54fff86e3fd325685c9e1c15b10" Oct 13 09:42:57 crc kubenswrapper[4685]: I1013 09:42:57.928219 4685 generic.go:334] "Generic (PLEG): container finished" podID="fe57c984-5668-4cdf-8faa-1c9057d82f32" containerID="ff08ea7e1627cab5d337438a0e8b47fb01c4bf5f99477d994d07febbd9106ed6" exitCode=0 Oct 13 09:42:57 crc kubenswrapper[4685]: I1013 09:42:57.928260 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mzp24/crc-debug-zq72k" event={"ID":"fe57c984-5668-4cdf-8faa-1c9057d82f32","Type":"ContainerDied","Data":"ff08ea7e1627cab5d337438a0e8b47fb01c4bf5f99477d994d07febbd9106ed6"} Oct 13 09:42:59 crc kubenswrapper[4685]: I1013 09:42:59.027676 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mzp24/crc-debug-zq72k" Oct 13 09:42:59 crc kubenswrapper[4685]: I1013 09:42:59.063745 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mzp24/crc-debug-zq72k"] Oct 13 09:42:59 crc kubenswrapper[4685]: I1013 09:42:59.072696 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mzp24/crc-debug-zq72k"] Oct 13 09:42:59 crc kubenswrapper[4685]: I1013 09:42:59.094890 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fe57c984-5668-4cdf-8faa-1c9057d82f32-host\") pod \"fe57c984-5668-4cdf-8faa-1c9057d82f32\" (UID: \"fe57c984-5668-4cdf-8faa-1c9057d82f32\") " Oct 13 09:42:59 crc kubenswrapper[4685]: I1013 09:42:59.095169 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fw5ml\" (UniqueName: \"kubernetes.io/projected/fe57c984-5668-4cdf-8faa-1c9057d82f32-kube-api-access-fw5ml\") pod \"fe57c984-5668-4cdf-8faa-1c9057d82f32\" (UID: \"fe57c984-5668-4cdf-8faa-1c9057d82f32\") " Oct 13 09:42:59 crc kubenswrapper[4685]: I1013 09:42:59.094986 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fe57c984-5668-4cdf-8faa-1c9057d82f32-host" (OuterVolumeSpecName: "host") pod "fe57c984-5668-4cdf-8faa-1c9057d82f32" (UID: "fe57c984-5668-4cdf-8faa-1c9057d82f32"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 09:42:59 crc kubenswrapper[4685]: I1013 09:42:59.095813 4685 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fe57c984-5668-4cdf-8faa-1c9057d82f32-host\") on node \"crc\" DevicePath \"\"" Oct 13 09:42:59 crc kubenswrapper[4685]: I1013 09:42:59.109198 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe57c984-5668-4cdf-8faa-1c9057d82f32-kube-api-access-fw5ml" (OuterVolumeSpecName: "kube-api-access-fw5ml") pod "fe57c984-5668-4cdf-8faa-1c9057d82f32" (UID: "fe57c984-5668-4cdf-8faa-1c9057d82f32"). InnerVolumeSpecName "kube-api-access-fw5ml". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:42:59 crc kubenswrapper[4685]: I1013 09:42:59.198003 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fw5ml\" (UniqueName: \"kubernetes.io/projected/fe57c984-5668-4cdf-8faa-1c9057d82f32-kube-api-access-fw5ml\") on node \"crc\" DevicePath \"\"" Oct 13 09:42:59 crc kubenswrapper[4685]: I1013 09:42:59.518713 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe57c984-5668-4cdf-8faa-1c9057d82f32" path="/var/lib/kubelet/pods/fe57c984-5668-4cdf-8faa-1c9057d82f32/volumes" Oct 13 09:42:59 crc kubenswrapper[4685]: I1013 09:42:59.953001 4685 scope.go:117] "RemoveContainer" containerID="ff08ea7e1627cab5d337438a0e8b47fb01c4bf5f99477d994d07febbd9106ed6" Oct 13 09:42:59 crc kubenswrapper[4685]: I1013 09:42:59.953121 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mzp24/crc-debug-zq72k" Oct 13 09:43:00 crc kubenswrapper[4685]: I1013 09:43:00.228593 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mzp24/crc-debug-pzbxk"] Oct 13 09:43:00 crc kubenswrapper[4685]: E1013 09:43:00.229053 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe57c984-5668-4cdf-8faa-1c9057d82f32" containerName="container-00" Oct 13 09:43:00 crc kubenswrapper[4685]: I1013 09:43:00.229066 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe57c984-5668-4cdf-8faa-1c9057d82f32" containerName="container-00" Oct 13 09:43:00 crc kubenswrapper[4685]: I1013 09:43:00.229254 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe57c984-5668-4cdf-8faa-1c9057d82f32" containerName="container-00" Oct 13 09:43:00 crc kubenswrapper[4685]: I1013 09:43:00.229833 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mzp24/crc-debug-pzbxk" Oct 13 09:43:00 crc kubenswrapper[4685]: I1013 09:43:00.323464 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/963de70d-fc2e-45a4-9a16-b8750a675e01-host\") pod \"crc-debug-pzbxk\" (UID: \"963de70d-fc2e-45a4-9a16-b8750a675e01\") " pod="openshift-must-gather-mzp24/crc-debug-pzbxk" Oct 13 09:43:00 crc kubenswrapper[4685]: I1013 09:43:00.323731 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xrmn\" (UniqueName: \"kubernetes.io/projected/963de70d-fc2e-45a4-9a16-b8750a675e01-kube-api-access-4xrmn\") pod \"crc-debug-pzbxk\" (UID: \"963de70d-fc2e-45a4-9a16-b8750a675e01\") " pod="openshift-must-gather-mzp24/crc-debug-pzbxk" Oct 13 09:43:00 crc kubenswrapper[4685]: I1013 09:43:00.425974 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/963de70d-fc2e-45a4-9a16-b8750a675e01-host\") pod \"crc-debug-pzbxk\" (UID: \"963de70d-fc2e-45a4-9a16-b8750a675e01\") " pod="openshift-must-gather-mzp24/crc-debug-pzbxk" Oct 13 09:43:00 crc kubenswrapper[4685]: I1013 09:43:00.426029 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xrmn\" (UniqueName: \"kubernetes.io/projected/963de70d-fc2e-45a4-9a16-b8750a675e01-kube-api-access-4xrmn\") pod \"crc-debug-pzbxk\" (UID: \"963de70d-fc2e-45a4-9a16-b8750a675e01\") " pod="openshift-must-gather-mzp24/crc-debug-pzbxk" Oct 13 09:43:00 crc kubenswrapper[4685]: I1013 09:43:00.426369 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/963de70d-fc2e-45a4-9a16-b8750a675e01-host\") pod \"crc-debug-pzbxk\" (UID: \"963de70d-fc2e-45a4-9a16-b8750a675e01\") " pod="openshift-must-gather-mzp24/crc-debug-pzbxk" Oct 13 09:43:00 crc kubenswrapper[4685]: I1013 09:43:00.464009 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xrmn\" (UniqueName: \"kubernetes.io/projected/963de70d-fc2e-45a4-9a16-b8750a675e01-kube-api-access-4xrmn\") pod \"crc-debug-pzbxk\" (UID: \"963de70d-fc2e-45a4-9a16-b8750a675e01\") " pod="openshift-must-gather-mzp24/crc-debug-pzbxk" Oct 13 09:43:00 crc kubenswrapper[4685]: I1013 09:43:00.545204 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mzp24/crc-debug-pzbxk" Oct 13 09:43:00 crc kubenswrapper[4685]: W1013 09:43:00.570135 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod963de70d_fc2e_45a4_9a16_b8750a675e01.slice/crio-2ba1b2e8bb98e69d045e108e0d853651700c69a6117eef9a36652936b1ec042c WatchSource:0}: Error finding container 2ba1b2e8bb98e69d045e108e0d853651700c69a6117eef9a36652936b1ec042c: Status 404 returned error can't find the container with id 2ba1b2e8bb98e69d045e108e0d853651700c69a6117eef9a36652936b1ec042c Oct 13 09:43:00 crc kubenswrapper[4685]: I1013 09:43:00.963398 4685 generic.go:334] "Generic (PLEG): container finished" podID="963de70d-fc2e-45a4-9a16-b8750a675e01" containerID="6ae2aaf4a37aad849c0ff972cf13a01eff0d2f2c4c9b9567b011ee39897b68ea" exitCode=0 Oct 13 09:43:00 crc kubenswrapper[4685]: I1013 09:43:00.963478 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mzp24/crc-debug-pzbxk" event={"ID":"963de70d-fc2e-45a4-9a16-b8750a675e01","Type":"ContainerDied","Data":"6ae2aaf4a37aad849c0ff972cf13a01eff0d2f2c4c9b9567b011ee39897b68ea"} Oct 13 09:43:00 crc kubenswrapper[4685]: I1013 09:43:00.963662 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mzp24/crc-debug-pzbxk" event={"ID":"963de70d-fc2e-45a4-9a16-b8750a675e01","Type":"ContainerStarted","Data":"2ba1b2e8bb98e69d045e108e0d853651700c69a6117eef9a36652936b1ec042c"} Oct 13 09:43:01 crc kubenswrapper[4685]: I1013 09:43:01.507995 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:43:01 crc kubenswrapper[4685]: I1013 09:43:01.976401 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" event={"ID":"b001d17a-1aea-44ba-86c5-ba6b312156c1","Type":"ContainerStarted","Data":"d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790"} Oct 13 09:43:01 crc kubenswrapper[4685]: I1013 09:43:01.977034 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:43:02 crc kubenswrapper[4685]: I1013 09:43:02.020065 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mzp24/crc-debug-pzbxk"] Oct 13 09:43:02 crc kubenswrapper[4685]: I1013 09:43:02.036567 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mzp24/crc-debug-pzbxk"] Oct 13 09:43:02 crc kubenswrapper[4685]: I1013 09:43:02.095350 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mzp24/crc-debug-pzbxk" Oct 13 09:43:02 crc kubenswrapper[4685]: I1013 09:43:02.155439 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/963de70d-fc2e-45a4-9a16-b8750a675e01-host\") pod \"963de70d-fc2e-45a4-9a16-b8750a675e01\" (UID: \"963de70d-fc2e-45a4-9a16-b8750a675e01\") " Oct 13 09:43:02 crc kubenswrapper[4685]: I1013 09:43:02.155619 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xrmn\" (UniqueName: \"kubernetes.io/projected/963de70d-fc2e-45a4-9a16-b8750a675e01-kube-api-access-4xrmn\") pod \"963de70d-fc2e-45a4-9a16-b8750a675e01\" (UID: \"963de70d-fc2e-45a4-9a16-b8750a675e01\") " Oct 13 09:43:02 crc kubenswrapper[4685]: I1013 09:43:02.157110 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/963de70d-fc2e-45a4-9a16-b8750a675e01-host" (OuterVolumeSpecName: "host") pod "963de70d-fc2e-45a4-9a16-b8750a675e01" (UID: "963de70d-fc2e-45a4-9a16-b8750a675e01"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 09:43:02 crc kubenswrapper[4685]: I1013 09:43:02.162506 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/963de70d-fc2e-45a4-9a16-b8750a675e01-kube-api-access-4xrmn" (OuterVolumeSpecName: "kube-api-access-4xrmn") pod "963de70d-fc2e-45a4-9a16-b8750a675e01" (UID: "963de70d-fc2e-45a4-9a16-b8750a675e01"). InnerVolumeSpecName "kube-api-access-4xrmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:43:02 crc kubenswrapper[4685]: I1013 09:43:02.257972 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xrmn\" (UniqueName: \"kubernetes.io/projected/963de70d-fc2e-45a4-9a16-b8750a675e01-kube-api-access-4xrmn\") on node \"crc\" DevicePath \"\"" Oct 13 09:43:02 crc kubenswrapper[4685]: I1013 09:43:02.258010 4685 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/963de70d-fc2e-45a4-9a16-b8750a675e01-host\") on node \"crc\" DevicePath \"\"" Oct 13 09:43:02 crc kubenswrapper[4685]: I1013 09:43:02.987325 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mzp24/crc-debug-pzbxk" Oct 13 09:43:02 crc kubenswrapper[4685]: I1013 09:43:02.987546 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ba1b2e8bb98e69d045e108e0d853651700c69a6117eef9a36652936b1ec042c" Oct 13 09:43:03 crc kubenswrapper[4685]: I1013 09:43:03.271710 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mzp24/crc-debug-mntk7"] Oct 13 09:43:03 crc kubenswrapper[4685]: E1013 09:43:03.272117 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="963de70d-fc2e-45a4-9a16-b8750a675e01" containerName="container-00" Oct 13 09:43:03 crc kubenswrapper[4685]: I1013 09:43:03.272132 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="963de70d-fc2e-45a4-9a16-b8750a675e01" containerName="container-00" Oct 13 09:43:03 crc kubenswrapper[4685]: I1013 09:43:03.272300 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="963de70d-fc2e-45a4-9a16-b8750a675e01" containerName="container-00" Oct 13 09:43:03 crc kubenswrapper[4685]: I1013 09:43:03.280285 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mzp24/crc-debug-mntk7" Oct 13 09:43:03 crc kubenswrapper[4685]: I1013 09:43:03.377933 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd-host\") pod \"crc-debug-mntk7\" (UID: \"9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd\") " pod="openshift-must-gather-mzp24/crc-debug-mntk7" Oct 13 09:43:03 crc kubenswrapper[4685]: I1013 09:43:03.378344 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsp2r\" (UniqueName: \"kubernetes.io/projected/9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd-kube-api-access-qsp2r\") pod \"crc-debug-mntk7\" (UID: \"9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd\") " pod="openshift-must-gather-mzp24/crc-debug-mntk7" Oct 13 09:43:03 crc kubenswrapper[4685]: I1013 09:43:03.479881 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd-host\") pod \"crc-debug-mntk7\" (UID: \"9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd\") " pod="openshift-must-gather-mzp24/crc-debug-mntk7" Oct 13 09:43:03 crc kubenswrapper[4685]: I1013 09:43:03.480114 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsp2r\" (UniqueName: \"kubernetes.io/projected/9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd-kube-api-access-qsp2r\") pod \"crc-debug-mntk7\" (UID: \"9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd\") " pod="openshift-must-gather-mzp24/crc-debug-mntk7" Oct 13 09:43:03 crc kubenswrapper[4685]: I1013 09:43:03.480012 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd-host\") pod \"crc-debug-mntk7\" (UID: \"9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd\") " pod="openshift-must-gather-mzp24/crc-debug-mntk7" Oct 13 09:43:03 crc kubenswrapper[4685]: I1013 09:43:03.496406 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsp2r\" (UniqueName: \"kubernetes.io/projected/9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd-kube-api-access-qsp2r\") pod \"crc-debug-mntk7\" (UID: \"9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd\") " pod="openshift-must-gather-mzp24/crc-debug-mntk7" Oct 13 09:43:03 crc kubenswrapper[4685]: I1013 09:43:03.513477 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="963de70d-fc2e-45a4-9a16-b8750a675e01" path="/var/lib/kubelet/pods/963de70d-fc2e-45a4-9a16-b8750a675e01/volumes" Oct 13 09:43:03 crc kubenswrapper[4685]: I1013 09:43:03.594750 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mzp24/crc-debug-mntk7" Oct 13 09:43:03 crc kubenswrapper[4685]: W1013 09:43:03.630615 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fa27e5f_ea8d_40e4_8d1c_462f3caba5cd.slice/crio-70ccab784a527b6015e2e1c1d542407897590187ea5698cb6edeb82f2c57f7d3 WatchSource:0}: Error finding container 70ccab784a527b6015e2e1c1d542407897590187ea5698cb6edeb82f2c57f7d3: Status 404 returned error can't find the container with id 70ccab784a527b6015e2e1c1d542407897590187ea5698cb6edeb82f2c57f7d3 Oct 13 09:43:03 crc kubenswrapper[4685]: I1013 09:43:03.996154 4685 generic.go:334] "Generic (PLEG): container finished" podID="9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd" containerID="3e4a409975c5e5dff317f3bb27ab9581c95a9a7fd62c463096809003c0b1a786" exitCode=0 Oct 13 09:43:03 crc kubenswrapper[4685]: I1013 09:43:03.996195 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mzp24/crc-debug-mntk7" event={"ID":"9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd","Type":"ContainerDied","Data":"3e4a409975c5e5dff317f3bb27ab9581c95a9a7fd62c463096809003c0b1a786"} Oct 13 09:43:03 crc kubenswrapper[4685]: I1013 09:43:03.996226 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mzp24/crc-debug-mntk7" event={"ID":"9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd","Type":"ContainerStarted","Data":"70ccab784a527b6015e2e1c1d542407897590187ea5698cb6edeb82f2c57f7d3"} Oct 13 09:43:04 crc kubenswrapper[4685]: I1013 09:43:04.032756 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mzp24/crc-debug-mntk7"] Oct 13 09:43:04 crc kubenswrapper[4685]: I1013 09:43:04.044119 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mzp24/crc-debug-mntk7"] Oct 13 09:43:05 crc kubenswrapper[4685]: I1013 09:43:05.095987 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mzp24/crc-debug-mntk7" Oct 13 09:43:05 crc kubenswrapper[4685]: I1013 09:43:05.215142 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd-host\") pod \"9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd\" (UID: \"9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd\") " Oct 13 09:43:05 crc kubenswrapper[4685]: I1013 09:43:05.215208 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsp2r\" (UniqueName: \"kubernetes.io/projected/9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd-kube-api-access-qsp2r\") pod \"9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd\" (UID: \"9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd\") " Oct 13 09:43:05 crc kubenswrapper[4685]: I1013 09:43:05.215255 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd-host" (OuterVolumeSpecName: "host") pod "9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd" (UID: "9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 09:43:05 crc kubenswrapper[4685]: I1013 09:43:05.215795 4685 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd-host\") on node \"crc\" DevicePath \"\"" Oct 13 09:43:05 crc kubenswrapper[4685]: I1013 09:43:05.233644 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd-kube-api-access-qsp2r" (OuterVolumeSpecName: "kube-api-access-qsp2r") pod "9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd" (UID: "9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd"). InnerVolumeSpecName "kube-api-access-qsp2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:43:05 crc kubenswrapper[4685]: I1013 09:43:05.317315 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsp2r\" (UniqueName: \"kubernetes.io/projected/9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd-kube-api-access-qsp2r\") on node \"crc\" DevicePath \"\"" Oct 13 09:43:05 crc kubenswrapper[4685]: I1013 09:43:05.516302 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd" path="/var/lib/kubelet/pods/9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd/volumes" Oct 13 09:43:06 crc kubenswrapper[4685]: I1013 09:43:06.013117 4685 scope.go:117] "RemoveContainer" containerID="3e4a409975c5e5dff317f3bb27ab9581c95a9a7fd62c463096809003c0b1a786" Oct 13 09:43:06 crc kubenswrapper[4685]: I1013 09:43:06.013290 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mzp24/crc-debug-mntk7" Oct 13 09:43:13 crc kubenswrapper[4685]: I1013 09:43:13.445625 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:43:13 crc kubenswrapper[4685]: I1013 09:43:13.933474 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9lnph"] Oct 13 09:43:13 crc kubenswrapper[4685]: E1013 09:43:13.933831 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd" containerName="container-00" Oct 13 09:43:13 crc kubenswrapper[4685]: I1013 09:43:13.933851 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd" containerName="container-00" Oct 13 09:43:13 crc kubenswrapper[4685]: I1013 09:43:13.934082 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fa27e5f-ea8d-40e4-8d1c-462f3caba5cd" containerName="container-00" Oct 13 09:43:13 crc kubenswrapper[4685]: I1013 09:43:13.935353 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9lnph" Oct 13 09:43:13 crc kubenswrapper[4685]: I1013 09:43:13.948597 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9lnph"] Oct 13 09:43:14 crc kubenswrapper[4685]: I1013 09:43:14.001272 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38e31de5-00b7-4619-b70b-5f9e9ce64d9b-catalog-content\") pod \"community-operators-9lnph\" (UID: \"38e31de5-00b7-4619-b70b-5f9e9ce64d9b\") " pod="openshift-marketplace/community-operators-9lnph" Oct 13 09:43:14 crc kubenswrapper[4685]: I1013 09:43:14.001527 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4f99v\" (UniqueName: \"kubernetes.io/projected/38e31de5-00b7-4619-b70b-5f9e9ce64d9b-kube-api-access-4f99v\") pod \"community-operators-9lnph\" (UID: \"38e31de5-00b7-4619-b70b-5f9e9ce64d9b\") " pod="openshift-marketplace/community-operators-9lnph" Oct 13 09:43:14 crc kubenswrapper[4685]: I1013 09:43:14.001588 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38e31de5-00b7-4619-b70b-5f9e9ce64d9b-utilities\") pod \"community-operators-9lnph\" (UID: \"38e31de5-00b7-4619-b70b-5f9e9ce64d9b\") " pod="openshift-marketplace/community-operators-9lnph" Oct 13 09:43:14 crc kubenswrapper[4685]: I1013 09:43:14.103089 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38e31de5-00b7-4619-b70b-5f9e9ce64d9b-catalog-content\") pod \"community-operators-9lnph\" (UID: \"38e31de5-00b7-4619-b70b-5f9e9ce64d9b\") " pod="openshift-marketplace/community-operators-9lnph" Oct 13 09:43:14 crc kubenswrapper[4685]: I1013 09:43:14.103145 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4f99v\" (UniqueName: \"kubernetes.io/projected/38e31de5-00b7-4619-b70b-5f9e9ce64d9b-kube-api-access-4f99v\") pod \"community-operators-9lnph\" (UID: \"38e31de5-00b7-4619-b70b-5f9e9ce64d9b\") " pod="openshift-marketplace/community-operators-9lnph" Oct 13 09:43:14 crc kubenswrapper[4685]: I1013 09:43:14.103224 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38e31de5-00b7-4619-b70b-5f9e9ce64d9b-utilities\") pod \"community-operators-9lnph\" (UID: \"38e31de5-00b7-4619-b70b-5f9e9ce64d9b\") " pod="openshift-marketplace/community-operators-9lnph" Oct 13 09:43:14 crc kubenswrapper[4685]: I1013 09:43:14.103669 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38e31de5-00b7-4619-b70b-5f9e9ce64d9b-catalog-content\") pod \"community-operators-9lnph\" (UID: \"38e31de5-00b7-4619-b70b-5f9e9ce64d9b\") " pod="openshift-marketplace/community-operators-9lnph" Oct 13 09:43:14 crc kubenswrapper[4685]: I1013 09:43:14.103729 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38e31de5-00b7-4619-b70b-5f9e9ce64d9b-utilities\") pod \"community-operators-9lnph\" (UID: \"38e31de5-00b7-4619-b70b-5f9e9ce64d9b\") " pod="openshift-marketplace/community-operators-9lnph" Oct 13 09:43:14 crc kubenswrapper[4685]: I1013 09:43:14.125352 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4f99v\" (UniqueName: \"kubernetes.io/projected/38e31de5-00b7-4619-b70b-5f9e9ce64d9b-kube-api-access-4f99v\") pod \"community-operators-9lnph\" (UID: \"38e31de5-00b7-4619-b70b-5f9e9ce64d9b\") " pod="openshift-marketplace/community-operators-9lnph" Oct 13 09:43:14 crc kubenswrapper[4685]: I1013 09:43:14.252794 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9lnph" Oct 13 09:43:14 crc kubenswrapper[4685]: I1013 09:43:14.811616 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9lnph"] Oct 13 09:43:14 crc kubenswrapper[4685]: W1013 09:43:14.817018 4685 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38e31de5_00b7_4619_b70b_5f9e9ce64d9b.slice/crio-79af9e5f604ef435adadc3753a6c14b30e51d582e29c7b4a95910c8e26c0ee88 WatchSource:0}: Error finding container 79af9e5f604ef435adadc3753a6c14b30e51d582e29c7b4a95910c8e26c0ee88: Status 404 returned error can't find the container with id 79af9e5f604ef435adadc3753a6c14b30e51d582e29c7b4a95910c8e26c0ee88 Oct 13 09:43:15 crc kubenswrapper[4685]: I1013 09:43:15.096762 4685 generic.go:334] "Generic (PLEG): container finished" podID="38e31de5-00b7-4619-b70b-5f9e9ce64d9b" containerID="e83ad32b7d55c846c21dac04542b9435cd4d5e0927938ce6f8b5b5c0beff22dc" exitCode=0 Oct 13 09:43:15 crc kubenswrapper[4685]: I1013 09:43:15.097024 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9lnph" event={"ID":"38e31de5-00b7-4619-b70b-5f9e9ce64d9b","Type":"ContainerDied","Data":"e83ad32b7d55c846c21dac04542b9435cd4d5e0927938ce6f8b5b5c0beff22dc"} Oct 13 09:43:15 crc kubenswrapper[4685]: I1013 09:43:15.097049 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9lnph" event={"ID":"38e31de5-00b7-4619-b70b-5f9e9ce64d9b","Type":"ContainerStarted","Data":"79af9e5f604ef435adadc3753a6c14b30e51d582e29c7b4a95910c8e26c0ee88"} Oct 13 09:43:16 crc kubenswrapper[4685]: I1013 09:43:16.107052 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9lnph" event={"ID":"38e31de5-00b7-4619-b70b-5f9e9ce64d9b","Type":"ContainerStarted","Data":"aef98a1ffd9e5c112770f0d4340f3ef2b6b8f22e73216b600b0e26f3da7cfe97"} Oct 13 09:43:17 crc kubenswrapper[4685]: I1013 09:43:17.116565 4685 generic.go:334] "Generic (PLEG): container finished" podID="38e31de5-00b7-4619-b70b-5f9e9ce64d9b" containerID="aef98a1ffd9e5c112770f0d4340f3ef2b6b8f22e73216b600b0e26f3da7cfe97" exitCode=0 Oct 13 09:43:17 crc kubenswrapper[4685]: I1013 09:43:17.116662 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9lnph" event={"ID":"38e31de5-00b7-4619-b70b-5f9e9ce64d9b","Type":"ContainerDied","Data":"aef98a1ffd9e5c112770f0d4340f3ef2b6b8f22e73216b600b0e26f3da7cfe97"} Oct 13 09:43:18 crc kubenswrapper[4685]: I1013 09:43:18.126902 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9lnph" event={"ID":"38e31de5-00b7-4619-b70b-5f9e9ce64d9b","Type":"ContainerStarted","Data":"6f298681bf34d74c649ad3e3ed452df63f5f21d7be7373a78db90274d40b7520"} Oct 13 09:43:18 crc kubenswrapper[4685]: I1013 09:43:18.147384 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9lnph" podStartSLOduration=2.719608509 podStartE2EDuration="5.147352335s" podCreationTimestamp="2025-10-13 09:43:13 +0000 UTC" firstStartedPulling="2025-10-13 09:43:15.099043936 +0000 UTC m=+3520.246919697" lastFinishedPulling="2025-10-13 09:43:17.526787762 +0000 UTC m=+3522.674663523" observedRunningTime="2025-10-13 09:43:18.146516183 +0000 UTC m=+3523.294391954" watchObservedRunningTime="2025-10-13 09:43:18.147352335 +0000 UTC m=+3523.295228096" Oct 13 09:43:24 crc kubenswrapper[4685]: I1013 09:43:24.253216 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9lnph" Oct 13 09:43:24 crc kubenswrapper[4685]: I1013 09:43:24.253803 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9lnph" Oct 13 09:43:24 crc kubenswrapper[4685]: I1013 09:43:24.338101 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9lnph" Oct 13 09:43:25 crc kubenswrapper[4685]: I1013 09:43:25.239059 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9lnph" Oct 13 09:43:25 crc kubenswrapper[4685]: I1013 09:43:25.290723 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9lnph"] Oct 13 09:43:27 crc kubenswrapper[4685]: I1013 09:43:27.195080 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9lnph" podUID="38e31de5-00b7-4619-b70b-5f9e9ce64d9b" containerName="registry-server" containerID="cri-o://6f298681bf34d74c649ad3e3ed452df63f5f21d7be7373a78db90274d40b7520" gracePeriod=2 Oct 13 09:43:27 crc kubenswrapper[4685]: I1013 09:43:27.726421 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9lnph" Oct 13 09:43:27 crc kubenswrapper[4685]: I1013 09:43:27.773697 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4f99v\" (UniqueName: \"kubernetes.io/projected/38e31de5-00b7-4619-b70b-5f9e9ce64d9b-kube-api-access-4f99v\") pod \"38e31de5-00b7-4619-b70b-5f9e9ce64d9b\" (UID: \"38e31de5-00b7-4619-b70b-5f9e9ce64d9b\") " Oct 13 09:43:27 crc kubenswrapper[4685]: I1013 09:43:27.773779 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38e31de5-00b7-4619-b70b-5f9e9ce64d9b-utilities\") pod \"38e31de5-00b7-4619-b70b-5f9e9ce64d9b\" (UID: \"38e31de5-00b7-4619-b70b-5f9e9ce64d9b\") " Oct 13 09:43:27 crc kubenswrapper[4685]: I1013 09:43:27.773858 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38e31de5-00b7-4619-b70b-5f9e9ce64d9b-catalog-content\") pod \"38e31de5-00b7-4619-b70b-5f9e9ce64d9b\" (UID: \"38e31de5-00b7-4619-b70b-5f9e9ce64d9b\") " Oct 13 09:43:27 crc kubenswrapper[4685]: I1013 09:43:27.776884 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38e31de5-00b7-4619-b70b-5f9e9ce64d9b-utilities" (OuterVolumeSpecName: "utilities") pod "38e31de5-00b7-4619-b70b-5f9e9ce64d9b" (UID: "38e31de5-00b7-4619-b70b-5f9e9ce64d9b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:43:27 crc kubenswrapper[4685]: I1013 09:43:27.780169 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38e31de5-00b7-4619-b70b-5f9e9ce64d9b-kube-api-access-4f99v" (OuterVolumeSpecName: "kube-api-access-4f99v") pod "38e31de5-00b7-4619-b70b-5f9e9ce64d9b" (UID: "38e31de5-00b7-4619-b70b-5f9e9ce64d9b"). InnerVolumeSpecName "kube-api-access-4f99v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:43:27 crc kubenswrapper[4685]: I1013 09:43:27.844941 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38e31de5-00b7-4619-b70b-5f9e9ce64d9b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38e31de5-00b7-4619-b70b-5f9e9ce64d9b" (UID: "38e31de5-00b7-4619-b70b-5f9e9ce64d9b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:43:27 crc kubenswrapper[4685]: I1013 09:43:27.875325 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4f99v\" (UniqueName: \"kubernetes.io/projected/38e31de5-00b7-4619-b70b-5f9e9ce64d9b-kube-api-access-4f99v\") on node \"crc\" DevicePath \"\"" Oct 13 09:43:27 crc kubenswrapper[4685]: I1013 09:43:27.875357 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38e31de5-00b7-4619-b70b-5f9e9ce64d9b-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 09:43:27 crc kubenswrapper[4685]: I1013 09:43:27.875365 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38e31de5-00b7-4619-b70b-5f9e9ce64d9b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 09:43:28 crc kubenswrapper[4685]: I1013 09:43:28.203939 4685 generic.go:334] "Generic (PLEG): container finished" podID="38e31de5-00b7-4619-b70b-5f9e9ce64d9b" containerID="6f298681bf34d74c649ad3e3ed452df63f5f21d7be7373a78db90274d40b7520" exitCode=0 Oct 13 09:43:28 crc kubenswrapper[4685]: I1013 09:43:28.203977 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9lnph" event={"ID":"38e31de5-00b7-4619-b70b-5f9e9ce64d9b","Type":"ContainerDied","Data":"6f298681bf34d74c649ad3e3ed452df63f5f21d7be7373a78db90274d40b7520"} Oct 13 09:43:28 crc kubenswrapper[4685]: I1013 09:43:28.204002 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9lnph" event={"ID":"38e31de5-00b7-4619-b70b-5f9e9ce64d9b","Type":"ContainerDied","Data":"79af9e5f604ef435adadc3753a6c14b30e51d582e29c7b4a95910c8e26c0ee88"} Oct 13 09:43:28 crc kubenswrapper[4685]: I1013 09:43:28.204017 4685 scope.go:117] "RemoveContainer" containerID="6f298681bf34d74c649ad3e3ed452df63f5f21d7be7373a78db90274d40b7520" Oct 13 09:43:28 crc kubenswrapper[4685]: I1013 09:43:28.204040 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9lnph" Oct 13 09:43:28 crc kubenswrapper[4685]: I1013 09:43:28.222362 4685 scope.go:117] "RemoveContainer" containerID="aef98a1ffd9e5c112770f0d4340f3ef2b6b8f22e73216b600b0e26f3da7cfe97" Oct 13 09:43:28 crc kubenswrapper[4685]: I1013 09:43:28.248672 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9lnph"] Oct 13 09:43:28 crc kubenswrapper[4685]: I1013 09:43:28.256227 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9lnph"] Oct 13 09:43:28 crc kubenswrapper[4685]: I1013 09:43:28.268366 4685 scope.go:117] "RemoveContainer" containerID="e83ad32b7d55c846c21dac04542b9435cd4d5e0927938ce6f8b5b5c0beff22dc" Oct 13 09:43:28 crc kubenswrapper[4685]: I1013 09:43:28.309570 4685 scope.go:117] "RemoveContainer" containerID="6f298681bf34d74c649ad3e3ed452df63f5f21d7be7373a78db90274d40b7520" Oct 13 09:43:28 crc kubenswrapper[4685]: E1013 09:43:28.310292 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f298681bf34d74c649ad3e3ed452df63f5f21d7be7373a78db90274d40b7520\": container with ID starting with 6f298681bf34d74c649ad3e3ed452df63f5f21d7be7373a78db90274d40b7520 not found: ID does not exist" containerID="6f298681bf34d74c649ad3e3ed452df63f5f21d7be7373a78db90274d40b7520" Oct 13 09:43:28 crc kubenswrapper[4685]: I1013 09:43:28.310321 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f298681bf34d74c649ad3e3ed452df63f5f21d7be7373a78db90274d40b7520"} err="failed to get container status \"6f298681bf34d74c649ad3e3ed452df63f5f21d7be7373a78db90274d40b7520\": rpc error: code = NotFound desc = could not find container \"6f298681bf34d74c649ad3e3ed452df63f5f21d7be7373a78db90274d40b7520\": container with ID starting with 6f298681bf34d74c649ad3e3ed452df63f5f21d7be7373a78db90274d40b7520 not found: ID does not exist" Oct 13 09:43:28 crc kubenswrapper[4685]: I1013 09:43:28.310342 4685 scope.go:117] "RemoveContainer" containerID="aef98a1ffd9e5c112770f0d4340f3ef2b6b8f22e73216b600b0e26f3da7cfe97" Oct 13 09:43:28 crc kubenswrapper[4685]: E1013 09:43:28.310715 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aef98a1ffd9e5c112770f0d4340f3ef2b6b8f22e73216b600b0e26f3da7cfe97\": container with ID starting with aef98a1ffd9e5c112770f0d4340f3ef2b6b8f22e73216b600b0e26f3da7cfe97 not found: ID does not exist" containerID="aef98a1ffd9e5c112770f0d4340f3ef2b6b8f22e73216b600b0e26f3da7cfe97" Oct 13 09:43:28 crc kubenswrapper[4685]: I1013 09:43:28.310809 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aef98a1ffd9e5c112770f0d4340f3ef2b6b8f22e73216b600b0e26f3da7cfe97"} err="failed to get container status \"aef98a1ffd9e5c112770f0d4340f3ef2b6b8f22e73216b600b0e26f3da7cfe97\": rpc error: code = NotFound desc = could not find container \"aef98a1ffd9e5c112770f0d4340f3ef2b6b8f22e73216b600b0e26f3da7cfe97\": container with ID starting with aef98a1ffd9e5c112770f0d4340f3ef2b6b8f22e73216b600b0e26f3da7cfe97 not found: ID does not exist" Oct 13 09:43:28 crc kubenswrapper[4685]: I1013 09:43:28.310900 4685 scope.go:117] "RemoveContainer" containerID="e83ad32b7d55c846c21dac04542b9435cd4d5e0927938ce6f8b5b5c0beff22dc" Oct 13 09:43:28 crc kubenswrapper[4685]: E1013 09:43:28.311253 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e83ad32b7d55c846c21dac04542b9435cd4d5e0927938ce6f8b5b5c0beff22dc\": container with ID starting with e83ad32b7d55c846c21dac04542b9435cd4d5e0927938ce6f8b5b5c0beff22dc not found: ID does not exist" containerID="e83ad32b7d55c846c21dac04542b9435cd4d5e0927938ce6f8b5b5c0beff22dc" Oct 13 09:43:28 crc kubenswrapper[4685]: I1013 09:43:28.311280 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e83ad32b7d55c846c21dac04542b9435cd4d5e0927938ce6f8b5b5c0beff22dc"} err="failed to get container status \"e83ad32b7d55c846c21dac04542b9435cd4d5e0927938ce6f8b5b5c0beff22dc\": rpc error: code = NotFound desc = could not find container \"e83ad32b7d55c846c21dac04542b9435cd4d5e0927938ce6f8b5b5c0beff22dc\": container with ID starting with e83ad32b7d55c846c21dac04542b9435cd4d5e0927938ce6f8b5b5c0beff22dc not found: ID does not exist" Oct 13 09:43:28 crc kubenswrapper[4685]: I1013 09:43:28.428499 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7cc9c86456-5hsl7_27d72d0b-473c-4983-be92-7b3eca7d5c45/barbican-api/0.log" Oct 13 09:43:28 crc kubenswrapper[4685]: I1013 09:43:28.505598 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7cc9c86456-5hsl7_27d72d0b-473c-4983-be92-7b3eca7d5c45/barbican-api-log/0.log" Oct 13 09:43:28 crc kubenswrapper[4685]: I1013 09:43:28.866376 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-58b69dbf78-cznmg_ad40cbfc-60f4-4ff0-9106-90a9941d5c10/barbican-keystone-listener/0.log" Oct 13 09:43:29 crc kubenswrapper[4685]: I1013 09:43:29.089392 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-58b69dbf78-cznmg_ad40cbfc-60f4-4ff0-9106-90a9941d5c10/barbican-keystone-listener-log/0.log" Oct 13 09:43:29 crc kubenswrapper[4685]: I1013 09:43:29.128077 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-f7b8c695-sw5hg_102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11/barbican-worker/0.log" Oct 13 09:43:29 crc kubenswrapper[4685]: I1013 09:43:29.130649 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-f7b8c695-sw5hg_102f7c1a-a920-4a3b-a9d1-fbb4c2aa5e11/barbican-worker-log/0.log" Oct 13 09:43:29 crc kubenswrapper[4685]: I1013 09:43:29.318388 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-zz5hl_5056abd4-630c-49fe-9ddf-1294f0f6b55a/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:43:29 crc kubenswrapper[4685]: I1013 09:43:29.385866 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5aa884f8-807f-407b-9ea0-ea0812fcd1a2/ceilometer-central-agent/0.log" Oct 13 09:43:29 crc kubenswrapper[4685]: I1013 09:43:29.496961 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5aa884f8-807f-407b-9ea0-ea0812fcd1a2/proxy-httpd/0.log" Oct 13 09:43:29 crc kubenswrapper[4685]: I1013 09:43:29.525264 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38e31de5-00b7-4619-b70b-5f9e9ce64d9b" path="/var/lib/kubelet/pods/38e31de5-00b7-4619-b70b-5f9e9ce64d9b/volumes" Oct 13 09:43:29 crc kubenswrapper[4685]: I1013 09:43:29.579167 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5aa884f8-807f-407b-9ea0-ea0812fcd1a2/ceilometer-notification-agent/0.log" Oct 13 09:43:29 crc kubenswrapper[4685]: I1013 09:43:29.588298 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5aa884f8-807f-407b-9ea0-ea0812fcd1a2/sg-core/0.log" Oct 13 09:43:29 crc kubenswrapper[4685]: I1013 09:43:29.773964 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f/cinder-api/0.log" Oct 13 09:43:29 crc kubenswrapper[4685]: I1013 09:43:29.907362 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9b177c68-df1a-4f9a-8a5f-0fd4749e5a2f/cinder-api-log/0.log" Oct 13 09:43:30 crc kubenswrapper[4685]: I1013 09:43:30.078637 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c60c26e9-8d95-47ec-9211-7053c42c471b/probe/0.log" Oct 13 09:43:30 crc kubenswrapper[4685]: I1013 09:43:30.086919 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c60c26e9-8d95-47ec-9211-7053c42c471b/cinder-scheduler/0.log" Oct 13 09:43:30 crc kubenswrapper[4685]: I1013 09:43:30.159144 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-w5pgs_f5d774ea-039f-4303-b889-8a6c79077453/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:43:30 crc kubenswrapper[4685]: I1013 09:43:30.404795 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-jk24b_c9f2aa20-f6e6-47e7-9244-36184ce2cf1c/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:43:30 crc kubenswrapper[4685]: I1013 09:43:30.409900 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-jhmms_2a07788a-cfaa-4c89-91ec-96999f31a6fd/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:43:30 crc kubenswrapper[4685]: I1013 09:43:30.609196 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b865b64bc-nsk84_908f2660-8000-46f8-8119-e43b00f79abb/init/0.log" Oct 13 09:43:30 crc kubenswrapper[4685]: I1013 09:43:30.750619 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b865b64bc-nsk84_908f2660-8000-46f8-8119-e43b00f79abb/init/0.log" Oct 13 09:43:30 crc kubenswrapper[4685]: I1013 09:43:30.868032 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b865b64bc-nsk84_908f2660-8000-46f8-8119-e43b00f79abb/dnsmasq-dns/0.log" Oct 13 09:43:30 crc kubenswrapper[4685]: I1013 09:43:30.928961 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-cwlnc_49edf7b8-3834-40c0-a038-e6c2388842da/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:43:31 crc kubenswrapper[4685]: I1013 09:43:31.046191 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_cf7cb050-76b4-4ee0-b71f-3ad58070fdc3/glance-httpd/0.log" Oct 13 09:43:31 crc kubenswrapper[4685]: I1013 09:43:31.090001 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_cf7cb050-76b4-4ee0-b71f-3ad58070fdc3/glance-log/0.log" Oct 13 09:43:31 crc kubenswrapper[4685]: I1013 09:43:31.273372 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_96ceb8f4-f3db-4f87-b5bf-27fd14076c1b/glance-httpd/0.log" Oct 13 09:43:31 crc kubenswrapper[4685]: I1013 09:43:31.299889 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_96ceb8f4-f3db-4f87-b5bf-27fd14076c1b/glance-log/0.log" Oct 13 09:43:31 crc kubenswrapper[4685]: I1013 09:43:31.440828 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-dbcf576b-msgft_46272601-40bd-43ec-b7a8-6824281972e0/horizon/0.log" Oct 13 09:43:31 crc kubenswrapper[4685]: I1013 09:43:31.626123 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-sc94f_e8ad7931-be11-41ff-9ca5-1ad434026e44/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:43:31 crc kubenswrapper[4685]: I1013 09:43:31.783744 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-dbcf576b-msgft_46272601-40bd-43ec-b7a8-6824281972e0/horizon-log/0.log" Oct 13 09:43:31 crc kubenswrapper[4685]: I1013 09:43:31.870125 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-s2r7z_5cb29fa8-e9f7-4d50-b65b-54f79a3a9de9/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:43:32 crc kubenswrapper[4685]: I1013 09:43:32.158589 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_6a895be7-8da1-4ba2-982a-bde7d7544581/kube-state-metrics/0.log" Oct 13 09:43:32 crc kubenswrapper[4685]: I1013 09:43:32.264547 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-79b7b9958d-7s7c7_869b9e65-4e24-4046-9b64-c1ad8d970be4/keystone-api/0.log" Oct 13 09:43:32 crc kubenswrapper[4685]: I1013 09:43:32.365885 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-5ckv2_1a2c2127-ae84-45bc-a364-caa9fe26133a/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:43:32 crc kubenswrapper[4685]: I1013 09:43:32.629795 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-786f745fcf-5snx2_4a0c8086-2865-4107-b424-b7820834b297/neutron-httpd/0.log" Oct 13 09:43:32 crc kubenswrapper[4685]: I1013 09:43:32.646202 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-786f745fcf-5snx2_4a0c8086-2865-4107-b424-b7820834b297/neutron-api/0.log" Oct 13 09:43:32 crc kubenswrapper[4685]: I1013 09:43:32.790841 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-6t5w5_0b2a22b2-f314-42af-95f0-f6213098f535/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:43:32 crc kubenswrapper[4685]: I1013 09:43:32.994806 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_fb5d643a-96c6-4fca-904d-3b4a71e9630c/memcached/0.log" Oct 13 09:43:33 crc kubenswrapper[4685]: I1013 09:43:33.130469 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_27a5deb0-f57d-4780-9426-7ff101dc91ed/nova-api-log/0.log" Oct 13 09:43:33 crc kubenswrapper[4685]: I1013 09:43:33.324135 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_95136327-c720-4660-9e60-adb85bd5afa2/nova-cell0-conductor-conductor/0.log" Oct 13 09:43:33 crc kubenswrapper[4685]: I1013 09:43:33.333497 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_27a5deb0-f57d-4780-9426-7ff101dc91ed/nova-api-api/0.log" Oct 13 09:43:33 crc kubenswrapper[4685]: I1013 09:43:33.397866 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_b4c0d942-9fdc-4f28-833c-1332cbe23c56/nova-cell1-conductor-conductor/0.log" Oct 13 09:43:33 crc kubenswrapper[4685]: I1013 09:43:33.627707 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-f2kmw_5e5b7b28-ee18-4b85-af9c-030cab7da493/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:43:33 crc kubenswrapper[4685]: I1013 09:43:33.671506 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_1039fca1-44b0-46d3-8905-daa12aadea65/nova-cell1-novncproxy-novncproxy/0.log" Oct 13 09:43:33 crc kubenswrapper[4685]: I1013 09:43:33.734400 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7ddd9484-88bb-4550-b9e7-399e001cfa29/nova-metadata-log/0.log" Oct 13 09:43:34 crc kubenswrapper[4685]: I1013 09:43:34.073186 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20/mysql-bootstrap/0.log" Oct 13 09:43:34 crc kubenswrapper[4685]: I1013 09:43:34.156642 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_c01b8af0-dd0a-4f1d-9e09-1e6ab0cf7779/nova-scheduler-scheduler/0.log" Oct 13 09:43:34 crc kubenswrapper[4685]: I1013 09:43:34.260190 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20/mysql-bootstrap/0.log" Oct 13 09:43:34 crc kubenswrapper[4685]: I1013 09:43:34.414864 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_66b058b9-a4a0-47a9-9a8b-b90e5e9e2b20/galera/0.log" Oct 13 09:43:34 crc kubenswrapper[4685]: I1013 09:43:34.470552 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f5482688-fe73-4196-ae4d-a306a1c31545/mysql-bootstrap/0.log" Oct 13 09:43:34 crc kubenswrapper[4685]: I1013 09:43:34.525276 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7ddd9484-88bb-4550-b9e7-399e001cfa29/nova-metadata-metadata/0.log" Oct 13 09:43:34 crc kubenswrapper[4685]: I1013 09:43:34.674224 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f5482688-fe73-4196-ae4d-a306a1c31545/mysql-bootstrap/0.log" Oct 13 09:43:34 crc kubenswrapper[4685]: I1013 09:43:34.697828 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f5482688-fe73-4196-ae4d-a306a1c31545/galera/0.log" Oct 13 09:43:34 crc kubenswrapper[4685]: I1013 09:43:34.759309 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_31c9baa6-8272-45f1-b8d4-6c89f710fe20/openstackclient/0.log" Oct 13 09:43:34 crc kubenswrapper[4685]: I1013 09:43:34.902241 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-t646z_85cf8998-48c3-49b9-9a68-d5765b785df4/openstack-network-exporter/0.log" Oct 13 09:43:35 crc kubenswrapper[4685]: I1013 09:43:35.021697 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gh7ls_fb7ad71b-d500-457d-81d4-831ba362a8b4/ovsdb-server-init/0.log" Oct 13 09:43:35 crc kubenswrapper[4685]: I1013 09:43:35.204387 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gh7ls_fb7ad71b-d500-457d-81d4-831ba362a8b4/ovsdb-server/0.log" Oct 13 09:43:35 crc kubenswrapper[4685]: I1013 09:43:35.260456 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gh7ls_fb7ad71b-d500-457d-81d4-831ba362a8b4/ovsdb-server-init/0.log" Oct 13 09:43:35 crc kubenswrapper[4685]: I1013 09:43:35.261824 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gh7ls_fb7ad71b-d500-457d-81d4-831ba362a8b4/ovs-vswitchd/0.log" Oct 13 09:43:35 crc kubenswrapper[4685]: I1013 09:43:35.273937 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-tb2c4_9dccff31-26fc-43c7-a679-01173d2f34a9/ovn-controller/0.log" Oct 13 09:43:35 crc kubenswrapper[4685]: I1013 09:43:35.484309 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-nwnvn_57483ac7-dfa4-4538-8605-11b9215432dc/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:43:35 crc kubenswrapper[4685]: I1013 09:43:35.487562 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_108cacab-202f-48f0-91f1-664e68d89e28/openstack-network-exporter/0.log" Oct 13 09:43:35 crc kubenswrapper[4685]: I1013 09:43:35.619424 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_108cacab-202f-48f0-91f1-664e68d89e28/ovn-northd/0.log" Oct 13 09:43:35 crc kubenswrapper[4685]: I1013 09:43:35.726145 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_7ffc3833-03fd-40ef-b247-4f1c512c2e8b/openstack-network-exporter/0.log" Oct 13 09:43:35 crc kubenswrapper[4685]: I1013 09:43:35.759742 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_7ffc3833-03fd-40ef-b247-4f1c512c2e8b/ovsdbserver-nb/0.log" Oct 13 09:43:35 crc kubenswrapper[4685]: I1013 09:43:35.864178 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_2a3319a1-2f17-43cd-9df1-0697ba10aff8/openstack-network-exporter/0.log" Oct 13 09:43:35 crc kubenswrapper[4685]: I1013 09:43:35.927312 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_2a3319a1-2f17-43cd-9df1-0697ba10aff8/ovsdbserver-sb/0.log" Oct 13 09:43:36 crc kubenswrapper[4685]: I1013 09:43:36.059777 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-55b747894d-xd2hx_9a194ffb-9cf0-4167-9c5b-c51bd79c42d7/placement-api/0.log" Oct 13 09:43:36 crc kubenswrapper[4685]: I1013 09:43:36.165857 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-55b747894d-xd2hx_9a194ffb-9cf0-4167-9c5b-c51bd79c42d7/placement-log/0.log" Oct 13 09:43:36 crc kubenswrapper[4685]: I1013 09:43:36.185821 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5f68bc3a-ebea-44a6-9b00-048e6afd1d09/setup-container/0.log" Oct 13 09:43:36 crc kubenswrapper[4685]: I1013 09:43:36.454385 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5f68bc3a-ebea-44a6-9b00-048e6afd1d09/setup-container/0.log" Oct 13 09:43:36 crc kubenswrapper[4685]: I1013 09:43:36.488487 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5f68bc3a-ebea-44a6-9b00-048e6afd1d09/rabbitmq/0.log" Oct 13 09:43:36 crc kubenswrapper[4685]: I1013 09:43:36.496241 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_27720755-e830-4eb4-b0e0-b5dfe9ceb253/setup-container/0.log" Oct 13 09:43:36 crc kubenswrapper[4685]: I1013 09:43:36.698285 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_27720755-e830-4eb4-b0e0-b5dfe9ceb253/rabbitmq/0.log" Oct 13 09:43:36 crc kubenswrapper[4685]: I1013 09:43:36.730864 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-wvstj_f6c7cd6b-3a10-490e-af5e-68a812d4b6f6/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:43:36 crc kubenswrapper[4685]: I1013 09:43:36.745642 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_27720755-e830-4eb4-b0e0-b5dfe9ceb253/setup-container/0.log" Oct 13 09:43:36 crc kubenswrapper[4685]: I1013 09:43:36.971201 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-gbjgj_cc857d68-fe9c-4e34-86c7-e26aca5432e6/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:43:37 crc kubenswrapper[4685]: I1013 09:43:37.042075 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-p2spf_82f29acf-d74b-434e-9aaf-7324cbc6c2d0/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:43:37 crc kubenswrapper[4685]: I1013 09:43:37.048313 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-m49qc_81fecbf5-ba69-4c64-b5c1-ae9f5e07d881/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:43:37 crc kubenswrapper[4685]: I1013 09:43:37.216780 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-xsmbj_73114199-2f57-451c-8b19-81ad0fbb98f5/ssh-known-hosts-edpm-deployment/0.log" Oct 13 09:43:37 crc kubenswrapper[4685]: I1013 09:43:37.417006 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-55864b6df5-xmtb6_86c3d3b4-cf99-46c5-b238-0efd9798f870/proxy-httpd/0.log" Oct 13 09:43:37 crc kubenswrapper[4685]: I1013 09:43:37.459787 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-55864b6df5-xmtb6_86c3d3b4-cf99-46c5-b238-0efd9798f870/proxy-server/0.log" Oct 13 09:43:37 crc kubenswrapper[4685]: I1013 09:43:37.510908 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-2qrrt_b884b6de-f048-4f14-b8b0-4775fa1d4cd1/swift-ring-rebalance/0.log" Oct 13 09:43:37 crc kubenswrapper[4685]: I1013 09:43:37.727226 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/account-auditor/0.log" Oct 13 09:43:37 crc kubenswrapper[4685]: I1013 09:43:37.751199 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/account-reaper/0.log" Oct 13 09:43:37 crc kubenswrapper[4685]: I1013 09:43:37.837954 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/account-replicator/0.log" Oct 13 09:43:37 crc kubenswrapper[4685]: I1013 09:43:37.851188 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/account-server/0.log" Oct 13 09:43:37 crc kubenswrapper[4685]: I1013 09:43:37.915204 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/container-auditor/0.log" Oct 13 09:43:37 crc kubenswrapper[4685]: I1013 09:43:37.957317 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/container-replicator/0.log" Oct 13 09:43:38 crc kubenswrapper[4685]: I1013 09:43:38.041368 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/object-auditor/0.log" Oct 13 09:43:38 crc kubenswrapper[4685]: I1013 09:43:38.052706 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/container-updater/0.log" Oct 13 09:43:38 crc kubenswrapper[4685]: I1013 09:43:38.075768 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/container-server/0.log" Oct 13 09:43:38 crc kubenswrapper[4685]: I1013 09:43:38.145623 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/object-expirer/0.log" Oct 13 09:43:38 crc kubenswrapper[4685]: I1013 09:43:38.233195 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/object-server/0.log" Oct 13 09:43:38 crc kubenswrapper[4685]: I1013 09:43:38.246066 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/object-replicator/0.log" Oct 13 09:43:38 crc kubenswrapper[4685]: I1013 09:43:38.259532 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/object-updater/0.log" Oct 13 09:43:38 crc kubenswrapper[4685]: I1013 09:43:38.331008 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/rsync/0.log" Oct 13 09:43:38 crc kubenswrapper[4685]: I1013 09:43:38.399344 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0791156b-11f2-43ad-b910-3e42b4d6670f/swift-recon-cron/0.log" Oct 13 09:43:38 crc kubenswrapper[4685]: I1013 09:43:38.484739 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-6xcbg_e955f1da-bde6-47d5-8cdf-fe0d7182c081/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:43:38 crc kubenswrapper[4685]: I1013 09:43:38.641746 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_804b452c-e552-4715-901c-061f1dc7db41/tempest-tests-tempest-tests-runner/0.log" Oct 13 09:43:38 crc kubenswrapper[4685]: I1013 09:43:38.765583 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_aed7cf9e-6686-44a8-91ef-a1e9613b0c65/test-operator-logs-container/0.log" Oct 13 09:43:38 crc kubenswrapper[4685]: I1013 09:43:38.790589 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-tpkbz_6535c66b-9311-4170-8fa5-c3d79b5cd7af/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 09:43:57 crc kubenswrapper[4685]: I1013 09:43:57.110639 4685 scope.go:117] "RemoveContainer" containerID="9de1401b38adcb6703c5ce010ce28506a4368a5fd5bbce0a69be23faedc86029" Oct 13 09:44:01 crc kubenswrapper[4685]: I1013 09:44:01.785292 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56_6c967b8a-ec86-413b-a25e-b81f3e9ac2dc/util/0.log" Oct 13 09:44:01 crc kubenswrapper[4685]: I1013 09:44:01.939993 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56_6c967b8a-ec86-413b-a25e-b81f3e9ac2dc/util/0.log" Oct 13 09:44:01 crc kubenswrapper[4685]: I1013 09:44:01.989064 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56_6c967b8a-ec86-413b-a25e-b81f3e9ac2dc/pull/0.log" Oct 13 09:44:02 crc kubenswrapper[4685]: I1013 09:44:02.005326 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56_6c967b8a-ec86-413b-a25e-b81f3e9ac2dc/pull/0.log" Oct 13 09:44:02 crc kubenswrapper[4685]: I1013 09:44:02.228387 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56_6c967b8a-ec86-413b-a25e-b81f3e9ac2dc/pull/0.log" Oct 13 09:44:02 crc kubenswrapper[4685]: I1013 09:44:02.229651 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56_6c967b8a-ec86-413b-a25e-b81f3e9ac2dc/util/0.log" Oct 13 09:44:02 crc kubenswrapper[4685]: I1013 09:44:02.268870 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4b9205a10f4547c56d6cf58e5b7964433bc06e04e38b3438c6efc32b7brlt56_6c967b8a-ec86-413b-a25e-b81f3e9ac2dc/extract/0.log" Oct 13 09:44:02 crc kubenswrapper[4685]: I1013 09:44:02.480427 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-5d7zf_e509e801-67de-4a55-bd22-cf3f73deca81/kube-rbac-proxy/0.log" Oct 13 09:44:02 crc kubenswrapper[4685]: I1013 09:44:02.515829 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-5d7zf_e509e801-67de-4a55-bd22-cf3f73deca81/manager/0.log" Oct 13 09:44:02 crc kubenswrapper[4685]: I1013 09:44:02.581954 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-qbl58_c92c1fba-b02b-4b6c-9570-b75ee60c5e86/kube-rbac-proxy/0.log" Oct 13 09:44:02 crc kubenswrapper[4685]: I1013 09:44:02.752749 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-qbl58_c92c1fba-b02b-4b6c-9570-b75ee60c5e86/manager/0.log" Oct 13 09:44:02 crc kubenswrapper[4685]: I1013 09:44:02.806033 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-hbnzh_92a042d7-669e-48d2-8d7e-8a8da4fc01eb/kube-rbac-proxy/0.log" Oct 13 09:44:02 crc kubenswrapper[4685]: I1013 09:44:02.814997 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-hbnzh_92a042d7-669e-48d2-8d7e-8a8da4fc01eb/manager/0.log" Oct 13 09:44:03 crc kubenswrapper[4685]: I1013 09:44:03.041424 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-zm9bd_54405c7c-61f7-41ea-ae0a-29128b51326c/kube-rbac-proxy/0.log" Oct 13 09:44:03 crc kubenswrapper[4685]: I1013 09:44:03.111532 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-zm9bd_54405c7c-61f7-41ea-ae0a-29128b51326c/manager/0.log" Oct 13 09:44:03 crc kubenswrapper[4685]: I1013 09:44:03.309657 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-mj7x8_6b61f732-3f39-459e-bec0-a6bdd5added4/kube-rbac-proxy/0.log" Oct 13 09:44:03 crc kubenswrapper[4685]: I1013 09:44:03.353696 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-mj7x8_6b61f732-3f39-459e-bec0-a6bdd5added4/manager/0.log" Oct 13 09:44:03 crc kubenswrapper[4685]: I1013 09:44:03.366006 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-sl5lv_eb08316b-7fd3-4d65-88e1-bbc91efcb7c7/kube-rbac-proxy/0.log" Oct 13 09:44:03 crc kubenswrapper[4685]: I1013 09:44:03.565728 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-sl5lv_eb08316b-7fd3-4d65-88e1-bbc91efcb7c7/manager/0.log" Oct 13 09:44:03 crc kubenswrapper[4685]: I1013 09:44:03.630048 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-2bcss_a3e83b7f-5017-44e7-b507-f46a2d3f5488/kube-rbac-proxy/0.log" Oct 13 09:44:03 crc kubenswrapper[4685]: I1013 09:44:03.746432 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-2bcss_a3e83b7f-5017-44e7-b507-f46a2d3f5488/manager/0.log" Oct 13 09:44:03 crc kubenswrapper[4685]: I1013 09:44:03.892827 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-xf9rd_9899b0d7-81a4-49f7-91cd-3c5aa72d49b2/kube-rbac-proxy/0.log" Oct 13 09:44:03 crc kubenswrapper[4685]: I1013 09:44:03.898561 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-xf9rd_9899b0d7-81a4-49f7-91cd-3c5aa72d49b2/manager/0.log" Oct 13 09:44:04 crc kubenswrapper[4685]: I1013 09:44:04.056572 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-f9d897d75-xtnqp_b001d17a-1aea-44ba-86c5-ba6b312156c1/kube-rbac-proxy/0.log" Oct 13 09:44:04 crc kubenswrapper[4685]: I1013 09:44:04.135326 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-f9d897d75-xtnqp_b001d17a-1aea-44ba-86c5-ba6b312156c1/manager/9.log" Oct 13 09:44:04 crc kubenswrapper[4685]: I1013 09:44:04.194816 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-f9d897d75-xtnqp_b001d17a-1aea-44ba-86c5-ba6b312156c1/manager/8.log" Oct 13 09:44:04 crc kubenswrapper[4685]: I1013 09:44:04.274853 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-q8dkt_12dc02ce-76d4-4376-ab6f-b2a75580cc4d/kube-rbac-proxy/0.log" Oct 13 09:44:04 crc kubenswrapper[4685]: I1013 09:44:04.344725 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-q8dkt_12dc02ce-76d4-4376-ab6f-b2a75580cc4d/manager/0.log" Oct 13 09:44:04 crc kubenswrapper[4685]: I1013 09:44:04.471408 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-j65jt_e297e3a5-6cd9-4314-b4dc-76ba4c25c6f5/kube-rbac-proxy/0.log" Oct 13 09:44:04 crc kubenswrapper[4685]: I1013 09:44:04.490271 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-j65jt_e297e3a5-6cd9-4314-b4dc-76ba4c25c6f5/manager/0.log" Oct 13 09:44:04 crc kubenswrapper[4685]: I1013 09:44:04.638497 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-cbwvq_de8c01ad-9e44-42f8-956c-0bb61165a222/kube-rbac-proxy/0.log" Oct 13 09:44:04 crc kubenswrapper[4685]: I1013 09:44:04.792349 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-cbwvq_de8c01ad-9e44-42f8-956c-0bb61165a222/manager/0.log" Oct 13 09:44:04 crc kubenswrapper[4685]: I1013 09:44:04.857047 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-ft77l_d363a932-2a5d-4082-841a-b5d677d0b1a1/kube-rbac-proxy/0.log" Oct 13 09:44:04 crc kubenswrapper[4685]: I1013 09:44:04.899810 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-ft77l_d363a932-2a5d-4082-841a-b5d677d0b1a1/manager/0.log" Oct 13 09:44:05 crc kubenswrapper[4685]: I1013 09:44:05.081661 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-b9jts_3e53276a-ccb3-40a1-b1e6-307cc335ff4d/kube-rbac-proxy/0.log" Oct 13 09:44:05 crc kubenswrapper[4685]: I1013 09:44:05.162564 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td_d4c8f24c-6e84-4931-8edb-504b184ea7b0/kube-rbac-proxy/0.log" Oct 13 09:44:05 crc kubenswrapper[4685]: I1013 09:44:05.181747 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-b9jts_3e53276a-ccb3-40a1-b1e6-307cc335ff4d/manager/0.log" Oct 13 09:44:05 crc kubenswrapper[4685]: I1013 09:44:05.316571 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757dnp5td_d4c8f24c-6e84-4931-8edb-504b184ea7b0/manager/0.log" Oct 13 09:44:05 crc kubenswrapper[4685]: I1013 09:44:05.415104 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-859d8f6f74-jwkhb_d2886500-f092-4cba-bd20-50483b96ceb3/kube-rbac-proxy/0.log" Oct 13 09:44:05 crc kubenswrapper[4685]: I1013 09:44:05.683796 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6c8cf6687-nbccj_c31c40db-9982-4692-91cc-26bf7b4ba509/kube-rbac-proxy/0.log" Oct 13 09:44:06 crc kubenswrapper[4685]: I1013 09:44:06.016026 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-t7wsk_3d6866fd-e05d-4533-9134-47a83469940b/registry-server/0.log" Oct 13 09:44:06 crc kubenswrapper[4685]: I1013 09:44:06.115385 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-7ktz6_a4e11db1-f272-469d-9c22-a649cdbcf95e/kube-rbac-proxy/0.log" Oct 13 09:44:06 crc kubenswrapper[4685]: I1013 09:44:06.208069 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6c8cf6687-nbccj_c31c40db-9982-4692-91cc-26bf7b4ba509/operator/0.log" Oct 13 09:44:06 crc kubenswrapper[4685]: I1013 09:44:06.346838 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-7ktz6_a4e11db1-f272-469d-9c22-a649cdbcf95e/manager/0.log" Oct 13 09:44:06 crc kubenswrapper[4685]: I1013 09:44:06.429139 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-fhw55_9239cb03-ddca-4542-b188-b89717a00f75/kube-rbac-proxy/0.log" Oct 13 09:44:06 crc kubenswrapper[4685]: I1013 09:44:06.516361 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-fhw55_9239cb03-ddca-4542-b188-b89717a00f75/manager/0.log" Oct 13 09:44:06 crc kubenswrapper[4685]: I1013 09:44:06.561933 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-859d8f6f74-jwkhb_d2886500-f092-4cba-bd20-50483b96ceb3/manager/0.log" Oct 13 09:44:06 crc kubenswrapper[4685]: I1013 09:44:06.640052 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-bgvk8_79c6993d-5a5b-4b38-b678-6c78e639d834/operator/0.log" Oct 13 09:44:06 crc kubenswrapper[4685]: I1013 09:44:06.717002 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-d6gn6_d639bb21-69ae-45c8-8a9c-aac17f57f8dd/kube-rbac-proxy/0.log" Oct 13 09:44:06 crc kubenswrapper[4685]: I1013 09:44:06.832689 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-d6gn6_d639bb21-69ae-45c8-8a9c-aac17f57f8dd/manager/0.log" Oct 13 09:44:06 crc kubenswrapper[4685]: I1013 09:44:06.914157 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-qnmxx_f1085de4-f906-4315-b8b5-a3ee4e7182c1/kube-rbac-proxy/0.log" Oct 13 09:44:06 crc kubenswrapper[4685]: I1013 09:44:06.980465 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-qnmxx_f1085de4-f906-4315-b8b5-a3ee4e7182c1/manager/0.log" Oct 13 09:44:07 crc kubenswrapper[4685]: I1013 09:44:07.047534 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-dd92k_0b4f3fea-3a89-4d28-89c0-436da959e36f/kube-rbac-proxy/0.log" Oct 13 09:44:07 crc kubenswrapper[4685]: I1013 09:44:07.138124 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-dd92k_0b4f3fea-3a89-4d28-89c0-436da959e36f/manager/0.log" Oct 13 09:44:07 crc kubenswrapper[4685]: I1013 09:44:07.180336 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-9wdvd_026b2615-8dc1-4ba7-83d1-1e21f4fa80d2/kube-rbac-proxy/0.log" Oct 13 09:44:07 crc kubenswrapper[4685]: I1013 09:44:07.235612 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-9wdvd_026b2615-8dc1-4ba7-83d1-1e21f4fa80d2/manager/0.log" Oct 13 09:44:22 crc kubenswrapper[4685]: I1013 09:44:22.979862 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:44:22 crc kubenswrapper[4685]: I1013 09:44:22.981079 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:44:23 crc kubenswrapper[4685]: I1013 09:44:23.202719 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-4zvdw_d869fa03-5196-4a23-a7d3-9bb709891678/control-plane-machine-set-operator/0.log" Oct 13 09:44:23 crc kubenswrapper[4685]: I1013 09:44:23.417537 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j6kt6_a185671e-0a3a-4a9e-a884-65b448b1e922/kube-rbac-proxy/0.log" Oct 13 09:44:23 crc kubenswrapper[4685]: I1013 09:44:23.465363 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j6kt6_a185671e-0a3a-4a9e-a884-65b448b1e922/machine-api-operator/0.log" Oct 13 09:44:25 crc kubenswrapper[4685]: I1013 09:44:25.680375 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5f6cg"] Oct 13 09:44:25 crc kubenswrapper[4685]: E1013 09:44:25.680812 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38e31de5-00b7-4619-b70b-5f9e9ce64d9b" containerName="registry-server" Oct 13 09:44:25 crc kubenswrapper[4685]: I1013 09:44:25.680826 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="38e31de5-00b7-4619-b70b-5f9e9ce64d9b" containerName="registry-server" Oct 13 09:44:25 crc kubenswrapper[4685]: E1013 09:44:25.680844 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38e31de5-00b7-4619-b70b-5f9e9ce64d9b" containerName="extract-content" Oct 13 09:44:25 crc kubenswrapper[4685]: I1013 09:44:25.680850 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="38e31de5-00b7-4619-b70b-5f9e9ce64d9b" containerName="extract-content" Oct 13 09:44:25 crc kubenswrapper[4685]: E1013 09:44:25.680873 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38e31de5-00b7-4619-b70b-5f9e9ce64d9b" containerName="extract-utilities" Oct 13 09:44:25 crc kubenswrapper[4685]: I1013 09:44:25.680880 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="38e31de5-00b7-4619-b70b-5f9e9ce64d9b" containerName="extract-utilities" Oct 13 09:44:25 crc kubenswrapper[4685]: I1013 09:44:25.681116 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="38e31de5-00b7-4619-b70b-5f9e9ce64d9b" containerName="registry-server" Oct 13 09:44:25 crc kubenswrapper[4685]: I1013 09:44:25.682489 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5f6cg" Oct 13 09:44:25 crc kubenswrapper[4685]: I1013 09:44:25.707746 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5f6cg"] Oct 13 09:44:25 crc kubenswrapper[4685]: I1013 09:44:25.856390 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bee2fc2-8ab7-447d-b993-714ed428ada1-utilities\") pod \"redhat-marketplace-5f6cg\" (UID: \"1bee2fc2-8ab7-447d-b993-714ed428ada1\") " pod="openshift-marketplace/redhat-marketplace-5f6cg" Oct 13 09:44:25 crc kubenswrapper[4685]: I1013 09:44:25.856576 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7ffs\" (UniqueName: \"kubernetes.io/projected/1bee2fc2-8ab7-447d-b993-714ed428ada1-kube-api-access-v7ffs\") pod \"redhat-marketplace-5f6cg\" (UID: \"1bee2fc2-8ab7-447d-b993-714ed428ada1\") " pod="openshift-marketplace/redhat-marketplace-5f6cg" Oct 13 09:44:25 crc kubenswrapper[4685]: I1013 09:44:25.856618 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bee2fc2-8ab7-447d-b993-714ed428ada1-catalog-content\") pod \"redhat-marketplace-5f6cg\" (UID: \"1bee2fc2-8ab7-447d-b993-714ed428ada1\") " pod="openshift-marketplace/redhat-marketplace-5f6cg" Oct 13 09:44:25 crc kubenswrapper[4685]: I1013 09:44:25.958114 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7ffs\" (UniqueName: \"kubernetes.io/projected/1bee2fc2-8ab7-447d-b993-714ed428ada1-kube-api-access-v7ffs\") pod \"redhat-marketplace-5f6cg\" (UID: \"1bee2fc2-8ab7-447d-b993-714ed428ada1\") " pod="openshift-marketplace/redhat-marketplace-5f6cg" Oct 13 09:44:25 crc kubenswrapper[4685]: I1013 09:44:25.958178 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bee2fc2-8ab7-447d-b993-714ed428ada1-catalog-content\") pod \"redhat-marketplace-5f6cg\" (UID: \"1bee2fc2-8ab7-447d-b993-714ed428ada1\") " pod="openshift-marketplace/redhat-marketplace-5f6cg" Oct 13 09:44:25 crc kubenswrapper[4685]: I1013 09:44:25.958229 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bee2fc2-8ab7-447d-b993-714ed428ada1-utilities\") pod \"redhat-marketplace-5f6cg\" (UID: \"1bee2fc2-8ab7-447d-b993-714ed428ada1\") " pod="openshift-marketplace/redhat-marketplace-5f6cg" Oct 13 09:44:25 crc kubenswrapper[4685]: I1013 09:44:25.958955 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bee2fc2-8ab7-447d-b993-714ed428ada1-utilities\") pod \"redhat-marketplace-5f6cg\" (UID: \"1bee2fc2-8ab7-447d-b993-714ed428ada1\") " pod="openshift-marketplace/redhat-marketplace-5f6cg" Oct 13 09:44:25 crc kubenswrapper[4685]: I1013 09:44:25.958997 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bee2fc2-8ab7-447d-b993-714ed428ada1-catalog-content\") pod \"redhat-marketplace-5f6cg\" (UID: \"1bee2fc2-8ab7-447d-b993-714ed428ada1\") " pod="openshift-marketplace/redhat-marketplace-5f6cg" Oct 13 09:44:25 crc kubenswrapper[4685]: I1013 09:44:25.985610 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7ffs\" (UniqueName: \"kubernetes.io/projected/1bee2fc2-8ab7-447d-b993-714ed428ada1-kube-api-access-v7ffs\") pod \"redhat-marketplace-5f6cg\" (UID: \"1bee2fc2-8ab7-447d-b993-714ed428ada1\") " pod="openshift-marketplace/redhat-marketplace-5f6cg" Oct 13 09:44:26 crc kubenswrapper[4685]: I1013 09:44:26.002636 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5f6cg" Oct 13 09:44:26 crc kubenswrapper[4685]: I1013 09:44:26.538887 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5f6cg"] Oct 13 09:44:26 crc kubenswrapper[4685]: I1013 09:44:26.683904 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5f6cg" event={"ID":"1bee2fc2-8ab7-447d-b993-714ed428ada1","Type":"ContainerStarted","Data":"0bfc3eaf0fc850b428f294a672cdd013e5b09939e9e25c074bc10ad5e744217f"} Oct 13 09:44:27 crc kubenswrapper[4685]: I1013 09:44:27.697482 4685 generic.go:334] "Generic (PLEG): container finished" podID="1bee2fc2-8ab7-447d-b993-714ed428ada1" containerID="9a6829b2b2b08fb28117d64b71fd97d0a6facf36fd32ce8c09841b7f3bed61a9" exitCode=0 Oct 13 09:44:27 crc kubenswrapper[4685]: I1013 09:44:27.699311 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5f6cg" event={"ID":"1bee2fc2-8ab7-447d-b993-714ed428ada1","Type":"ContainerDied","Data":"9a6829b2b2b08fb28117d64b71fd97d0a6facf36fd32ce8c09841b7f3bed61a9"} Oct 13 09:44:29 crc kubenswrapper[4685]: I1013 09:44:29.717445 4685 generic.go:334] "Generic (PLEG): container finished" podID="1bee2fc2-8ab7-447d-b993-714ed428ada1" containerID="33958facab75664da02e250da7fb5859539709c59e4da59945471662344a1ebb" exitCode=0 Oct 13 09:44:29 crc kubenswrapper[4685]: I1013 09:44:29.717611 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5f6cg" event={"ID":"1bee2fc2-8ab7-447d-b993-714ed428ada1","Type":"ContainerDied","Data":"33958facab75664da02e250da7fb5859539709c59e4da59945471662344a1ebb"} Oct 13 09:44:30 crc kubenswrapper[4685]: I1013 09:44:30.736403 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5f6cg" event={"ID":"1bee2fc2-8ab7-447d-b993-714ed428ada1","Type":"ContainerStarted","Data":"996167be0524d242b70c3a2fe45cc7d4bf4664568e767262032a7b92f8b03db0"} Oct 13 09:44:30 crc kubenswrapper[4685]: I1013 09:44:30.774129 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5f6cg" podStartSLOduration=3.317774115 podStartE2EDuration="5.774112266s" podCreationTimestamp="2025-10-13 09:44:25 +0000 UTC" firstStartedPulling="2025-10-13 09:44:27.703385779 +0000 UTC m=+3592.851261540" lastFinishedPulling="2025-10-13 09:44:30.15972393 +0000 UTC m=+3595.307599691" observedRunningTime="2025-10-13 09:44:30.767172108 +0000 UTC m=+3595.915047889" watchObservedRunningTime="2025-10-13 09:44:30.774112266 +0000 UTC m=+3595.921988017" Oct 13 09:44:36 crc kubenswrapper[4685]: I1013 09:44:36.003874 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5f6cg" Oct 13 09:44:36 crc kubenswrapper[4685]: I1013 09:44:36.005467 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5f6cg" Oct 13 09:44:36 crc kubenswrapper[4685]: I1013 09:44:36.062498 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5f6cg" Oct 13 09:44:36 crc kubenswrapper[4685]: I1013 09:44:36.870876 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5f6cg" Oct 13 09:44:36 crc kubenswrapper[4685]: I1013 09:44:36.918684 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5f6cg"] Oct 13 09:44:37 crc kubenswrapper[4685]: I1013 09:44:37.431839 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-k77gp_1fec02c3-da30-45db-9004-73214b0d5a33/cert-manager-controller/0.log" Oct 13 09:44:37 crc kubenswrapper[4685]: I1013 09:44:37.726597 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-8l97z_d1799c36-72e8-4563-b4df-0115cd0e1108/cert-manager-webhook/0.log" Oct 13 09:44:37 crc kubenswrapper[4685]: I1013 09:44:37.737084 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-pw6jw_17313b92-8f5e-45c2-986c-d86cea130b7e/cert-manager-cainjector/0.log" Oct 13 09:44:38 crc kubenswrapper[4685]: I1013 09:44:38.806863 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5f6cg" podUID="1bee2fc2-8ab7-447d-b993-714ed428ada1" containerName="registry-server" containerID="cri-o://996167be0524d242b70c3a2fe45cc7d4bf4664568e767262032a7b92f8b03db0" gracePeriod=2 Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.348071 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5f6cg" Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.490294 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bee2fc2-8ab7-447d-b993-714ed428ada1-catalog-content\") pod \"1bee2fc2-8ab7-447d-b993-714ed428ada1\" (UID: \"1bee2fc2-8ab7-447d-b993-714ed428ada1\") " Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.490499 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bee2fc2-8ab7-447d-b993-714ed428ada1-utilities\") pod \"1bee2fc2-8ab7-447d-b993-714ed428ada1\" (UID: \"1bee2fc2-8ab7-447d-b993-714ed428ada1\") " Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.490631 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7ffs\" (UniqueName: \"kubernetes.io/projected/1bee2fc2-8ab7-447d-b993-714ed428ada1-kube-api-access-v7ffs\") pod \"1bee2fc2-8ab7-447d-b993-714ed428ada1\" (UID: \"1bee2fc2-8ab7-447d-b993-714ed428ada1\") " Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.492010 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bee2fc2-8ab7-447d-b993-714ed428ada1-utilities" (OuterVolumeSpecName: "utilities") pod "1bee2fc2-8ab7-447d-b993-714ed428ada1" (UID: "1bee2fc2-8ab7-447d-b993-714ed428ada1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.509113 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bee2fc2-8ab7-447d-b993-714ed428ada1-kube-api-access-v7ffs" (OuterVolumeSpecName: "kube-api-access-v7ffs") pod "1bee2fc2-8ab7-447d-b993-714ed428ada1" (UID: "1bee2fc2-8ab7-447d-b993-714ed428ada1"). InnerVolumeSpecName "kube-api-access-v7ffs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.517157 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bee2fc2-8ab7-447d-b993-714ed428ada1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1bee2fc2-8ab7-447d-b993-714ed428ada1" (UID: "1bee2fc2-8ab7-447d-b993-714ed428ada1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.593373 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bee2fc2-8ab7-447d-b993-714ed428ada1-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.593634 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7ffs\" (UniqueName: \"kubernetes.io/projected/1bee2fc2-8ab7-447d-b993-714ed428ada1-kube-api-access-v7ffs\") on node \"crc\" DevicePath \"\"" Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.593713 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bee2fc2-8ab7-447d-b993-714ed428ada1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.835351 4685 generic.go:334] "Generic (PLEG): container finished" podID="1bee2fc2-8ab7-447d-b993-714ed428ada1" containerID="996167be0524d242b70c3a2fe45cc7d4bf4664568e767262032a7b92f8b03db0" exitCode=0 Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.835394 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5f6cg" event={"ID":"1bee2fc2-8ab7-447d-b993-714ed428ada1","Type":"ContainerDied","Data":"996167be0524d242b70c3a2fe45cc7d4bf4664568e767262032a7b92f8b03db0"} Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.835420 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5f6cg" event={"ID":"1bee2fc2-8ab7-447d-b993-714ed428ada1","Type":"ContainerDied","Data":"0bfc3eaf0fc850b428f294a672cdd013e5b09939e9e25c074bc10ad5e744217f"} Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.835447 4685 scope.go:117] "RemoveContainer" containerID="996167be0524d242b70c3a2fe45cc7d4bf4664568e767262032a7b92f8b03db0" Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.835513 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5f6cg" Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.872144 4685 scope.go:117] "RemoveContainer" containerID="33958facab75664da02e250da7fb5859539709c59e4da59945471662344a1ebb" Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.874419 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5f6cg"] Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.882217 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5f6cg"] Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.910924 4685 scope.go:117] "RemoveContainer" containerID="9a6829b2b2b08fb28117d64b71fd97d0a6facf36fd32ce8c09841b7f3bed61a9" Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.960532 4685 scope.go:117] "RemoveContainer" containerID="996167be0524d242b70c3a2fe45cc7d4bf4664568e767262032a7b92f8b03db0" Oct 13 09:44:39 crc kubenswrapper[4685]: E1013 09:44:39.961111 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"996167be0524d242b70c3a2fe45cc7d4bf4664568e767262032a7b92f8b03db0\": container with ID starting with 996167be0524d242b70c3a2fe45cc7d4bf4664568e767262032a7b92f8b03db0 not found: ID does not exist" containerID="996167be0524d242b70c3a2fe45cc7d4bf4664568e767262032a7b92f8b03db0" Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.961152 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"996167be0524d242b70c3a2fe45cc7d4bf4664568e767262032a7b92f8b03db0"} err="failed to get container status \"996167be0524d242b70c3a2fe45cc7d4bf4664568e767262032a7b92f8b03db0\": rpc error: code = NotFound desc = could not find container \"996167be0524d242b70c3a2fe45cc7d4bf4664568e767262032a7b92f8b03db0\": container with ID starting with 996167be0524d242b70c3a2fe45cc7d4bf4664568e767262032a7b92f8b03db0 not found: ID does not exist" Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.961173 4685 scope.go:117] "RemoveContainer" containerID="33958facab75664da02e250da7fb5859539709c59e4da59945471662344a1ebb" Oct 13 09:44:39 crc kubenswrapper[4685]: E1013 09:44:39.961516 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33958facab75664da02e250da7fb5859539709c59e4da59945471662344a1ebb\": container with ID starting with 33958facab75664da02e250da7fb5859539709c59e4da59945471662344a1ebb not found: ID does not exist" containerID="33958facab75664da02e250da7fb5859539709c59e4da59945471662344a1ebb" Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.961562 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33958facab75664da02e250da7fb5859539709c59e4da59945471662344a1ebb"} err="failed to get container status \"33958facab75664da02e250da7fb5859539709c59e4da59945471662344a1ebb\": rpc error: code = NotFound desc = could not find container \"33958facab75664da02e250da7fb5859539709c59e4da59945471662344a1ebb\": container with ID starting with 33958facab75664da02e250da7fb5859539709c59e4da59945471662344a1ebb not found: ID does not exist" Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.961596 4685 scope.go:117] "RemoveContainer" containerID="9a6829b2b2b08fb28117d64b71fd97d0a6facf36fd32ce8c09841b7f3bed61a9" Oct 13 09:44:39 crc kubenswrapper[4685]: E1013 09:44:39.961928 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a6829b2b2b08fb28117d64b71fd97d0a6facf36fd32ce8c09841b7f3bed61a9\": container with ID starting with 9a6829b2b2b08fb28117d64b71fd97d0a6facf36fd32ce8c09841b7f3bed61a9 not found: ID does not exist" containerID="9a6829b2b2b08fb28117d64b71fd97d0a6facf36fd32ce8c09841b7f3bed61a9" Oct 13 09:44:39 crc kubenswrapper[4685]: I1013 09:44:39.961948 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a6829b2b2b08fb28117d64b71fd97d0a6facf36fd32ce8c09841b7f3bed61a9"} err="failed to get container status \"9a6829b2b2b08fb28117d64b71fd97d0a6facf36fd32ce8c09841b7f3bed61a9\": rpc error: code = NotFound desc = could not find container \"9a6829b2b2b08fb28117d64b71fd97d0a6facf36fd32ce8c09841b7f3bed61a9\": container with ID starting with 9a6829b2b2b08fb28117d64b71fd97d0a6facf36fd32ce8c09841b7f3bed61a9 not found: ID does not exist" Oct 13 09:44:41 crc kubenswrapper[4685]: I1013 09:44:41.522977 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bee2fc2-8ab7-447d-b993-714ed428ada1" path="/var/lib/kubelet/pods/1bee2fc2-8ab7-447d-b993-714ed428ada1/volumes" Oct 13 09:44:51 crc kubenswrapper[4685]: I1013 09:44:51.210684 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-zscjx_1c0f8ea9-312a-479f-be32-33028d5d6651/nmstate-console-plugin/0.log" Oct 13 09:44:51 crc kubenswrapper[4685]: I1013 09:44:51.451826 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-5cc6j_ed739611-c4b8-42d8-9c8f-3382d0347c39/nmstate-handler/0.log" Oct 13 09:44:51 crc kubenswrapper[4685]: I1013 09:44:51.515465 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-bgw9w_4cdd7309-1689-4b3a-8165-bda0f92016c9/kube-rbac-proxy/0.log" Oct 13 09:44:51 crc kubenswrapper[4685]: I1013 09:44:51.640440 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-bgw9w_4cdd7309-1689-4b3a-8165-bda0f92016c9/nmstate-metrics/0.log" Oct 13 09:44:51 crc kubenswrapper[4685]: I1013 09:44:51.781317 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-7l6n7_4d407a70-cfdb-427d-94a8-c975db126733/nmstate-operator/0.log" Oct 13 09:44:51 crc kubenswrapper[4685]: I1013 09:44:51.909730 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-66rzd_ac20f0e3-6bff-43fa-93d5-6447bf249314/nmstate-webhook/0.log" Oct 13 09:44:52 crc kubenswrapper[4685]: I1013 09:44:52.980205 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:44:52 crc kubenswrapper[4685]: I1013 09:44:52.980264 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:45:00 crc kubenswrapper[4685]: I1013 09:45:00.172110 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339145-bd8n5"] Oct 13 09:45:00 crc kubenswrapper[4685]: E1013 09:45:00.173014 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bee2fc2-8ab7-447d-b993-714ed428ada1" containerName="registry-server" Oct 13 09:45:00 crc kubenswrapper[4685]: I1013 09:45:00.173028 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bee2fc2-8ab7-447d-b993-714ed428ada1" containerName="registry-server" Oct 13 09:45:00 crc kubenswrapper[4685]: E1013 09:45:00.173060 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bee2fc2-8ab7-447d-b993-714ed428ada1" containerName="extract-content" Oct 13 09:45:00 crc kubenswrapper[4685]: I1013 09:45:00.173066 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bee2fc2-8ab7-447d-b993-714ed428ada1" containerName="extract-content" Oct 13 09:45:00 crc kubenswrapper[4685]: E1013 09:45:00.173087 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bee2fc2-8ab7-447d-b993-714ed428ada1" containerName="extract-utilities" Oct 13 09:45:00 crc kubenswrapper[4685]: I1013 09:45:00.173095 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bee2fc2-8ab7-447d-b993-714ed428ada1" containerName="extract-utilities" Oct 13 09:45:00 crc kubenswrapper[4685]: I1013 09:45:00.173309 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bee2fc2-8ab7-447d-b993-714ed428ada1" containerName="registry-server" Oct 13 09:45:00 crc kubenswrapper[4685]: I1013 09:45:00.174030 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339145-bd8n5" Oct 13 09:45:00 crc kubenswrapper[4685]: I1013 09:45:00.175945 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 13 09:45:00 crc kubenswrapper[4685]: I1013 09:45:00.176147 4685 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 13 09:45:00 crc kubenswrapper[4685]: I1013 09:45:00.188673 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339145-bd8n5"] Oct 13 09:45:00 crc kubenswrapper[4685]: I1013 09:45:00.281651 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1de46b3c-7e5e-42bd-acf3-98498991275f-secret-volume\") pod \"collect-profiles-29339145-bd8n5\" (UID: \"1de46b3c-7e5e-42bd-acf3-98498991275f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339145-bd8n5" Oct 13 09:45:00 crc kubenswrapper[4685]: I1013 09:45:00.281998 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1de46b3c-7e5e-42bd-acf3-98498991275f-config-volume\") pod \"collect-profiles-29339145-bd8n5\" (UID: \"1de46b3c-7e5e-42bd-acf3-98498991275f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339145-bd8n5" Oct 13 09:45:00 crc kubenswrapper[4685]: I1013 09:45:00.282121 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8ltr\" (UniqueName: \"kubernetes.io/projected/1de46b3c-7e5e-42bd-acf3-98498991275f-kube-api-access-f8ltr\") pod \"collect-profiles-29339145-bd8n5\" (UID: \"1de46b3c-7e5e-42bd-acf3-98498991275f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339145-bd8n5" Oct 13 09:45:00 crc kubenswrapper[4685]: I1013 09:45:00.384307 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1de46b3c-7e5e-42bd-acf3-98498991275f-config-volume\") pod \"collect-profiles-29339145-bd8n5\" (UID: \"1de46b3c-7e5e-42bd-acf3-98498991275f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339145-bd8n5" Oct 13 09:45:00 crc kubenswrapper[4685]: I1013 09:45:00.384732 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8ltr\" (UniqueName: \"kubernetes.io/projected/1de46b3c-7e5e-42bd-acf3-98498991275f-kube-api-access-f8ltr\") pod \"collect-profiles-29339145-bd8n5\" (UID: \"1de46b3c-7e5e-42bd-acf3-98498991275f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339145-bd8n5" Oct 13 09:45:00 crc kubenswrapper[4685]: I1013 09:45:00.384784 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1de46b3c-7e5e-42bd-acf3-98498991275f-secret-volume\") pod \"collect-profiles-29339145-bd8n5\" (UID: \"1de46b3c-7e5e-42bd-acf3-98498991275f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339145-bd8n5" Oct 13 09:45:00 crc kubenswrapper[4685]: I1013 09:45:00.386860 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1de46b3c-7e5e-42bd-acf3-98498991275f-config-volume\") pod \"collect-profiles-29339145-bd8n5\" (UID: \"1de46b3c-7e5e-42bd-acf3-98498991275f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339145-bd8n5" Oct 13 09:45:00 crc kubenswrapper[4685]: I1013 09:45:00.400960 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1de46b3c-7e5e-42bd-acf3-98498991275f-secret-volume\") pod \"collect-profiles-29339145-bd8n5\" (UID: \"1de46b3c-7e5e-42bd-acf3-98498991275f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339145-bd8n5" Oct 13 09:45:00 crc kubenswrapper[4685]: I1013 09:45:00.406650 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8ltr\" (UniqueName: \"kubernetes.io/projected/1de46b3c-7e5e-42bd-acf3-98498991275f-kube-api-access-f8ltr\") pod \"collect-profiles-29339145-bd8n5\" (UID: \"1de46b3c-7e5e-42bd-acf3-98498991275f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339145-bd8n5" Oct 13 09:45:00 crc kubenswrapper[4685]: I1013 09:45:00.498159 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339145-bd8n5" Oct 13 09:45:01 crc kubenswrapper[4685]: I1013 09:45:01.071267 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339145-bd8n5"] Oct 13 09:45:02 crc kubenswrapper[4685]: I1013 09:45:02.017157 4685 generic.go:334] "Generic (PLEG): container finished" podID="1de46b3c-7e5e-42bd-acf3-98498991275f" containerID="2d485957ae2e96f775922906ab210b621b18e2c1c54214d2b86d9cca80232365" exitCode=0 Oct 13 09:45:02 crc kubenswrapper[4685]: I1013 09:45:02.017204 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339145-bd8n5" event={"ID":"1de46b3c-7e5e-42bd-acf3-98498991275f","Type":"ContainerDied","Data":"2d485957ae2e96f775922906ab210b621b18e2c1c54214d2b86d9cca80232365"} Oct 13 09:45:02 crc kubenswrapper[4685]: I1013 09:45:02.017512 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339145-bd8n5" event={"ID":"1de46b3c-7e5e-42bd-acf3-98498991275f","Type":"ContainerStarted","Data":"240fbf24a90732a0a186a3a98f8c2eb9fc1a264fe206cca5306e522610c8567b"} Oct 13 09:45:03 crc kubenswrapper[4685]: I1013 09:45:03.367820 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339145-bd8n5" Oct 13 09:45:03 crc kubenswrapper[4685]: I1013 09:45:03.439219 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1de46b3c-7e5e-42bd-acf3-98498991275f-secret-volume\") pod \"1de46b3c-7e5e-42bd-acf3-98498991275f\" (UID: \"1de46b3c-7e5e-42bd-acf3-98498991275f\") " Oct 13 09:45:03 crc kubenswrapper[4685]: I1013 09:45:03.439407 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8ltr\" (UniqueName: \"kubernetes.io/projected/1de46b3c-7e5e-42bd-acf3-98498991275f-kube-api-access-f8ltr\") pod \"1de46b3c-7e5e-42bd-acf3-98498991275f\" (UID: \"1de46b3c-7e5e-42bd-acf3-98498991275f\") " Oct 13 09:45:03 crc kubenswrapper[4685]: I1013 09:45:03.439475 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1de46b3c-7e5e-42bd-acf3-98498991275f-config-volume\") pod \"1de46b3c-7e5e-42bd-acf3-98498991275f\" (UID: \"1de46b3c-7e5e-42bd-acf3-98498991275f\") " Oct 13 09:45:03 crc kubenswrapper[4685]: I1013 09:45:03.440061 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1de46b3c-7e5e-42bd-acf3-98498991275f-config-volume" (OuterVolumeSpecName: "config-volume") pod "1de46b3c-7e5e-42bd-acf3-98498991275f" (UID: "1de46b3c-7e5e-42bd-acf3-98498991275f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 09:45:03 crc kubenswrapper[4685]: I1013 09:45:03.445355 4685 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1de46b3c-7e5e-42bd-acf3-98498991275f-config-volume\") on node \"crc\" DevicePath \"\"" Oct 13 09:45:03 crc kubenswrapper[4685]: I1013 09:45:03.448109 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1de46b3c-7e5e-42bd-acf3-98498991275f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1de46b3c-7e5e-42bd-acf3-98498991275f" (UID: "1de46b3c-7e5e-42bd-acf3-98498991275f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 09:45:03 crc kubenswrapper[4685]: I1013 09:45:03.448791 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1de46b3c-7e5e-42bd-acf3-98498991275f-kube-api-access-f8ltr" (OuterVolumeSpecName: "kube-api-access-f8ltr") pod "1de46b3c-7e5e-42bd-acf3-98498991275f" (UID: "1de46b3c-7e5e-42bd-acf3-98498991275f"). InnerVolumeSpecName "kube-api-access-f8ltr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:45:03 crc kubenswrapper[4685]: I1013 09:45:03.548930 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8ltr\" (UniqueName: \"kubernetes.io/projected/1de46b3c-7e5e-42bd-acf3-98498991275f-kube-api-access-f8ltr\") on node \"crc\" DevicePath \"\"" Oct 13 09:45:03 crc kubenswrapper[4685]: I1013 09:45:03.548975 4685 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1de46b3c-7e5e-42bd-acf3-98498991275f-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 13 09:45:04 crc kubenswrapper[4685]: I1013 09:45:04.035831 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339145-bd8n5" event={"ID":"1de46b3c-7e5e-42bd-acf3-98498991275f","Type":"ContainerDied","Data":"240fbf24a90732a0a186a3a98f8c2eb9fc1a264fe206cca5306e522610c8567b"} Oct 13 09:45:04 crc kubenswrapper[4685]: I1013 09:45:04.036206 4685 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="240fbf24a90732a0a186a3a98f8c2eb9fc1a264fe206cca5306e522610c8567b" Oct 13 09:45:04 crc kubenswrapper[4685]: I1013 09:45:04.035900 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339145-bd8n5" Oct 13 09:45:04 crc kubenswrapper[4685]: I1013 09:45:04.449809 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339100-h77rw"] Oct 13 09:45:04 crc kubenswrapper[4685]: I1013 09:45:04.456957 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339100-h77rw"] Oct 13 09:45:05 crc kubenswrapper[4685]: I1013 09:45:05.531456 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edbf1439-6678-4cef-a3ec-ba71bbe1bba6" path="/var/lib/kubelet/pods/edbf1439-6678-4cef-a3ec-ba71bbe1bba6/volumes" Oct 13 09:45:09 crc kubenswrapper[4685]: I1013 09:45:09.183530 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-fmsz9_34bb5061-30c8-49f5-8dca-a411b3f128b2/kube-rbac-proxy/0.log" Oct 13 09:45:09 crc kubenswrapper[4685]: I1013 09:45:09.184652 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-fmsz9_34bb5061-30c8-49f5-8dca-a411b3f128b2/controller/0.log" Oct 13 09:45:09 crc kubenswrapper[4685]: I1013 09:45:09.333324 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-frr-files/0.log" Oct 13 09:45:09 crc kubenswrapper[4685]: I1013 09:45:09.612120 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-reloader/0.log" Oct 13 09:45:09 crc kubenswrapper[4685]: I1013 09:45:09.627524 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-frr-files/0.log" Oct 13 09:45:09 crc kubenswrapper[4685]: I1013 09:45:09.630249 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-metrics/0.log" Oct 13 09:45:09 crc kubenswrapper[4685]: I1013 09:45:09.673123 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-reloader/0.log" Oct 13 09:45:09 crc kubenswrapper[4685]: I1013 09:45:09.946039 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-frr-files/0.log" Oct 13 09:45:09 crc kubenswrapper[4685]: I1013 09:45:09.986070 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-reloader/0.log" Oct 13 09:45:10 crc kubenswrapper[4685]: I1013 09:45:10.010162 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-metrics/0.log" Oct 13 09:45:10 crc kubenswrapper[4685]: I1013 09:45:10.053068 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-metrics/0.log" Oct 13 09:45:10 crc kubenswrapper[4685]: I1013 09:45:10.216069 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-reloader/0.log" Oct 13 09:45:10 crc kubenswrapper[4685]: I1013 09:45:10.223358 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-metrics/0.log" Oct 13 09:45:10 crc kubenswrapper[4685]: I1013 09:45:10.248035 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/cp-frr-files/0.log" Oct 13 09:45:10 crc kubenswrapper[4685]: I1013 09:45:10.275802 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/controller/0.log" Oct 13 09:45:10 crc kubenswrapper[4685]: I1013 09:45:10.479825 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/frr-metrics/0.log" Oct 13 09:45:10 crc kubenswrapper[4685]: I1013 09:45:10.533089 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/kube-rbac-proxy/0.log" Oct 13 09:45:10 crc kubenswrapper[4685]: I1013 09:45:10.540291 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/kube-rbac-proxy-frr/0.log" Oct 13 09:45:11 crc kubenswrapper[4685]: I1013 09:45:11.061207 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/reloader/0.log" Oct 13 09:45:11 crc kubenswrapper[4685]: I1013 09:45:11.094967 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-fkd9z_9b097f6c-ec81-4342-b4c1-63520267ba08/frr-k8s-webhook-server/0.log" Oct 13 09:45:11 crc kubenswrapper[4685]: I1013 09:45:11.325867 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-759fc95d5d-79ckz_b3d2fe67-48db-4dbc-869d-d0c13b18ec8a/manager/0.log" Oct 13 09:45:11 crc kubenswrapper[4685]: I1013 09:45:11.487572 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ggf8f_33a86974-4a79-42b8-8bf9-cf3992b9b54d/frr/0.log" Oct 13 09:45:11 crc kubenswrapper[4685]: I1013 09:45:11.584860 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6f5fffbd86-pw7w6_ce259f59-820d-41e4-8ef7-3b977664b7ea/webhook-server/0.log" Oct 13 09:45:11 crc kubenswrapper[4685]: I1013 09:45:11.658831 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-ddb2s_55ccd123-f671-4230-b2d5-e6ffb265429a/kube-rbac-proxy/0.log" Oct 13 09:45:12 crc kubenswrapper[4685]: I1013 09:45:12.013538 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-ddb2s_55ccd123-f671-4230-b2d5-e6ffb265429a/speaker/0.log" Oct 13 09:45:22 crc kubenswrapper[4685]: I1013 09:45:22.980349 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:45:22 crc kubenswrapper[4685]: I1013 09:45:22.981001 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:45:22 crc kubenswrapper[4685]: I1013 09:45:22.981050 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 09:45:22 crc kubenswrapper[4685]: I1013 09:45:22.981800 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a75682cfcb35d15ac8d83500d0d93801124ebd38b4b8e7001d309f70d4ce10ad"} pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 09:45:22 crc kubenswrapper[4685]: I1013 09:45:22.981864 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" containerID="cri-o://a75682cfcb35d15ac8d83500d0d93801124ebd38b4b8e7001d309f70d4ce10ad" gracePeriod=600 Oct 13 09:45:23 crc kubenswrapper[4685]: I1013 09:45:23.200314 4685 generic.go:334] "Generic (PLEG): container finished" podID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerID="a75682cfcb35d15ac8d83500d0d93801124ebd38b4b8e7001d309f70d4ce10ad" exitCode=0 Oct 13 09:45:23 crc kubenswrapper[4685]: I1013 09:45:23.200399 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerDied","Data":"a75682cfcb35d15ac8d83500d0d93801124ebd38b4b8e7001d309f70d4ce10ad"} Oct 13 09:45:23 crc kubenswrapper[4685]: I1013 09:45:23.200603 4685 scope.go:117] "RemoveContainer" containerID="dab01494d9d22c71a3d16e18eb05025316e8ee43c34b6e7e58ed7eaa8c953eb6" Oct 13 09:45:24 crc kubenswrapper[4685]: I1013 09:45:24.211211 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerStarted","Data":"78e305b74ad025a53109325e9b144a07a1448d8fcb50659d2a199e0a7538b084"} Oct 13 09:45:25 crc kubenswrapper[4685]: I1013 09:45:25.870331 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z_624e992c-aa90-41ff-a746-adf45015b81a/util/0.log" Oct 13 09:45:26 crc kubenswrapper[4685]: I1013 09:45:26.118436 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z_624e992c-aa90-41ff-a746-adf45015b81a/util/0.log" Oct 13 09:45:26 crc kubenswrapper[4685]: I1013 09:45:26.121644 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z_624e992c-aa90-41ff-a746-adf45015b81a/pull/0.log" Oct 13 09:45:26 crc kubenswrapper[4685]: I1013 09:45:26.207486 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z_624e992c-aa90-41ff-a746-adf45015b81a/pull/0.log" Oct 13 09:45:26 crc kubenswrapper[4685]: I1013 09:45:26.394056 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z_624e992c-aa90-41ff-a746-adf45015b81a/util/0.log" Oct 13 09:45:26 crc kubenswrapper[4685]: I1013 09:45:26.406561 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z_624e992c-aa90-41ff-a746-adf45015b81a/extract/0.log" Oct 13 09:45:26 crc kubenswrapper[4685]: I1013 09:45:26.727200 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24rz7z_624e992c-aa90-41ff-a746-adf45015b81a/pull/0.log" Oct 13 09:45:26 crc kubenswrapper[4685]: I1013 09:45:26.862523 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-flhsp_4f92f2fa-5809-4485-b5e1-90e6ebf8efd0/extract-utilities/0.log" Oct 13 09:45:27 crc kubenswrapper[4685]: I1013 09:45:27.091482 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-flhsp_4f92f2fa-5809-4485-b5e1-90e6ebf8efd0/extract-utilities/0.log" Oct 13 09:45:27 crc kubenswrapper[4685]: I1013 09:45:27.107079 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-flhsp_4f92f2fa-5809-4485-b5e1-90e6ebf8efd0/extract-content/0.log" Oct 13 09:45:27 crc kubenswrapper[4685]: I1013 09:45:27.112179 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-flhsp_4f92f2fa-5809-4485-b5e1-90e6ebf8efd0/extract-content/0.log" Oct 13 09:45:27 crc kubenswrapper[4685]: I1013 09:45:27.343376 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-flhsp_4f92f2fa-5809-4485-b5e1-90e6ebf8efd0/extract-content/0.log" Oct 13 09:45:27 crc kubenswrapper[4685]: I1013 09:45:27.388270 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-flhsp_4f92f2fa-5809-4485-b5e1-90e6ebf8efd0/extract-utilities/0.log" Oct 13 09:45:27 crc kubenswrapper[4685]: I1013 09:45:27.793899 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9z22p_d4dc80bb-b601-4804-80f0-e3f97493b377/extract-utilities/0.log" Oct 13 09:45:27 crc kubenswrapper[4685]: I1013 09:45:27.814113 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-flhsp_4f92f2fa-5809-4485-b5e1-90e6ebf8efd0/registry-server/0.log" Oct 13 09:45:27 crc kubenswrapper[4685]: I1013 09:45:27.982227 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9z22p_d4dc80bb-b601-4804-80f0-e3f97493b377/extract-utilities/0.log" Oct 13 09:45:28 crc kubenswrapper[4685]: I1013 09:45:28.028779 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9z22p_d4dc80bb-b601-4804-80f0-e3f97493b377/extract-content/0.log" Oct 13 09:45:28 crc kubenswrapper[4685]: I1013 09:45:28.063225 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9z22p_d4dc80bb-b601-4804-80f0-e3f97493b377/extract-content/0.log" Oct 13 09:45:28 crc kubenswrapper[4685]: I1013 09:45:28.223191 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9z22p_d4dc80bb-b601-4804-80f0-e3f97493b377/extract-content/0.log" Oct 13 09:45:28 crc kubenswrapper[4685]: I1013 09:45:28.251855 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9z22p_d4dc80bb-b601-4804-80f0-e3f97493b377/extract-utilities/0.log" Oct 13 09:45:28 crc kubenswrapper[4685]: I1013 09:45:28.514391 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc_207d7e9f-5cfb-4476-962f-8f6b960d32d2/util/0.log" Oct 13 09:45:28 crc kubenswrapper[4685]: I1013 09:45:28.850212 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9z22p_d4dc80bb-b601-4804-80f0-e3f97493b377/registry-server/0.log" Oct 13 09:45:28 crc kubenswrapper[4685]: I1013 09:45:28.945560 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc_207d7e9f-5cfb-4476-962f-8f6b960d32d2/util/0.log" Oct 13 09:45:28 crc kubenswrapper[4685]: I1013 09:45:28.957881 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc_207d7e9f-5cfb-4476-962f-8f6b960d32d2/pull/0.log" Oct 13 09:45:28 crc kubenswrapper[4685]: I1013 09:45:28.993316 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc_207d7e9f-5cfb-4476-962f-8f6b960d32d2/pull/0.log" Oct 13 09:45:29 crc kubenswrapper[4685]: I1013 09:45:29.083149 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc_207d7e9f-5cfb-4476-962f-8f6b960d32d2/util/0.log" Oct 13 09:45:29 crc kubenswrapper[4685]: I1013 09:45:29.169081 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc_207d7e9f-5cfb-4476-962f-8f6b960d32d2/pull/0.log" Oct 13 09:45:29 crc kubenswrapper[4685]: I1013 09:45:29.239902 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cth9rc_207d7e9f-5cfb-4476-962f-8f6b960d32d2/extract/0.log" Oct 13 09:45:29 crc kubenswrapper[4685]: I1013 09:45:29.285072 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-5pssn_30028087-7eef-4beb-8011-304f2dce2e01/marketplace-operator/0.log" Oct 13 09:45:29 crc kubenswrapper[4685]: I1013 09:45:29.479459 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pfbfb_396234a0-0c33-45f1-9e9a-a52c6e69e7c0/extract-utilities/0.log" Oct 13 09:45:29 crc kubenswrapper[4685]: I1013 09:45:29.751147 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pfbfb_396234a0-0c33-45f1-9e9a-a52c6e69e7c0/extract-utilities/0.log" Oct 13 09:45:29 crc kubenswrapper[4685]: I1013 09:45:29.801835 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pfbfb_396234a0-0c33-45f1-9e9a-a52c6e69e7c0/extract-content/0.log" Oct 13 09:45:29 crc kubenswrapper[4685]: I1013 09:45:29.812158 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pfbfb_396234a0-0c33-45f1-9e9a-a52c6e69e7c0/extract-content/0.log" Oct 13 09:45:30 crc kubenswrapper[4685]: I1013 09:45:30.063156 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pfbfb_396234a0-0c33-45f1-9e9a-a52c6e69e7c0/extract-content/0.log" Oct 13 09:45:30 crc kubenswrapper[4685]: I1013 09:45:30.093034 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pfbfb_396234a0-0c33-45f1-9e9a-a52c6e69e7c0/extract-utilities/0.log" Oct 13 09:45:30 crc kubenswrapper[4685]: I1013 09:45:30.214117 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pfbfb_396234a0-0c33-45f1-9e9a-a52c6e69e7c0/registry-server/0.log" Oct 13 09:45:30 crc kubenswrapper[4685]: I1013 09:45:30.294120 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t9v8k_9adb7fcf-2d97-4c07-85cf-86cef286dc69/extract-utilities/0.log" Oct 13 09:45:30 crc kubenswrapper[4685]: I1013 09:45:30.529155 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t9v8k_9adb7fcf-2d97-4c07-85cf-86cef286dc69/extract-content/0.log" Oct 13 09:45:30 crc kubenswrapper[4685]: I1013 09:45:30.531394 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t9v8k_9adb7fcf-2d97-4c07-85cf-86cef286dc69/extract-utilities/0.log" Oct 13 09:45:30 crc kubenswrapper[4685]: I1013 09:45:30.532883 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t9v8k_9adb7fcf-2d97-4c07-85cf-86cef286dc69/extract-content/0.log" Oct 13 09:45:30 crc kubenswrapper[4685]: I1013 09:45:30.695537 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t9v8k_9adb7fcf-2d97-4c07-85cf-86cef286dc69/extract-content/0.log" Oct 13 09:45:30 crc kubenswrapper[4685]: I1013 09:45:30.733805 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t9v8k_9adb7fcf-2d97-4c07-85cf-86cef286dc69/extract-utilities/0.log" Oct 13 09:45:31 crc kubenswrapper[4685]: I1013 09:45:31.178450 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t9v8k_9adb7fcf-2d97-4c07-85cf-86cef286dc69/registry-server/0.log" Oct 13 09:45:34 crc kubenswrapper[4685]: I1013 09:45:34.292149 4685 generic.go:334] "Generic (PLEG): container finished" podID="b001d17a-1aea-44ba-86c5-ba6b312156c1" containerID="d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790" exitCode=1 Oct 13 09:45:34 crc kubenswrapper[4685]: I1013 09:45:34.292382 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" event={"ID":"b001d17a-1aea-44ba-86c5-ba6b312156c1","Type":"ContainerDied","Data":"d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790"} Oct 13 09:45:34 crc kubenswrapper[4685]: I1013 09:45:34.292423 4685 scope.go:117] "RemoveContainer" containerID="096d96774ac46edf5a07f5a2fe4bcb2c03073a6925946c054180e9fe0ffe915a" Oct 13 09:45:34 crc kubenswrapper[4685]: I1013 09:45:34.293567 4685 scope.go:117] "RemoveContainer" containerID="d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790" Oct 13 09:45:34 crc kubenswrapper[4685]: E1013 09:45:34.293863 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:45:43 crc kubenswrapper[4685]: I1013 09:45:43.442796 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:45:43 crc kubenswrapper[4685]: I1013 09:45:43.443391 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" Oct 13 09:45:43 crc kubenswrapper[4685]: I1013 09:45:43.444196 4685 scope.go:117] "RemoveContainer" containerID="d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790" Oct 13 09:45:43 crc kubenswrapper[4685]: E1013 09:45:43.444542 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:45:55 crc kubenswrapper[4685]: I1013 09:45:55.512580 4685 scope.go:117] "RemoveContainer" containerID="d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790" Oct 13 09:45:55 crc kubenswrapper[4685]: E1013 09:45:55.513240 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:45:57 crc kubenswrapper[4685]: I1013 09:45:57.270202 4685 scope.go:117] "RemoveContainer" containerID="0afedaf401d57b82882990da486231433d3b6bb69c611ef47e1b0e99b494766f" Oct 13 09:46:06 crc kubenswrapper[4685]: I1013 09:46:06.503244 4685 scope.go:117] "RemoveContainer" containerID="d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790" Oct 13 09:46:06 crc kubenswrapper[4685]: E1013 09:46:06.503913 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:46:20 crc kubenswrapper[4685]: I1013 09:46:20.503251 4685 scope.go:117] "RemoveContainer" containerID="d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790" Oct 13 09:46:20 crc kubenswrapper[4685]: E1013 09:46:20.504435 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:46:33 crc kubenswrapper[4685]: I1013 09:46:33.503139 4685 scope.go:117] "RemoveContainer" containerID="d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790" Oct 13 09:46:33 crc kubenswrapper[4685]: E1013 09:46:33.504307 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:46:46 crc kubenswrapper[4685]: I1013 09:46:46.503406 4685 scope.go:117] "RemoveContainer" containerID="d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790" Oct 13 09:46:46 crc kubenswrapper[4685]: E1013 09:46:46.505958 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:46:58 crc kubenswrapper[4685]: I1013 09:46:58.503471 4685 scope.go:117] "RemoveContainer" containerID="d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790" Oct 13 09:46:58 crc kubenswrapper[4685]: E1013 09:46:58.504149 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:47:09 crc kubenswrapper[4685]: I1013 09:47:09.505559 4685 scope.go:117] "RemoveContainer" containerID="d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790" Oct 13 09:47:09 crc kubenswrapper[4685]: E1013 09:47:09.506397 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:47:10 crc kubenswrapper[4685]: I1013 09:47:10.404822 4685 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 13 09:47:21 crc kubenswrapper[4685]: I1013 09:47:21.503668 4685 scope.go:117] "RemoveContainer" containerID="d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790" Oct 13 09:47:21 crc kubenswrapper[4685]: E1013 09:47:21.504862 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:47:22 crc kubenswrapper[4685]: E1013 09:47:22.828117 4685 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5fc0e95d_4f82_4d0f_a745_001486489f27.slice/crio-de9ab10066e4fb70062591e9cee78610af292f02240bc5a8f175ce5d92ab3aa3.scope\": RecentStats: unable to find data in memory cache]" Oct 13 09:47:23 crc kubenswrapper[4685]: I1013 09:47:23.334691 4685 generic.go:334] "Generic (PLEG): container finished" podID="5fc0e95d-4f82-4d0f-a745-001486489f27" containerID="de9ab10066e4fb70062591e9cee78610af292f02240bc5a8f175ce5d92ab3aa3" exitCode=0 Oct 13 09:47:23 crc kubenswrapper[4685]: I1013 09:47:23.334746 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mzp24/must-gather-d64pf" event={"ID":"5fc0e95d-4f82-4d0f-a745-001486489f27","Type":"ContainerDied","Data":"de9ab10066e4fb70062591e9cee78610af292f02240bc5a8f175ce5d92ab3aa3"} Oct 13 09:47:23 crc kubenswrapper[4685]: I1013 09:47:23.336226 4685 scope.go:117] "RemoveContainer" containerID="de9ab10066e4fb70062591e9cee78610af292f02240bc5a8f175ce5d92ab3aa3" Oct 13 09:47:23 crc kubenswrapper[4685]: I1013 09:47:23.868192 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mzp24_must-gather-d64pf_5fc0e95d-4f82-4d0f-a745-001486489f27/gather/0.log" Oct 13 09:47:28 crc kubenswrapper[4685]: I1013 09:47:28.763181 4685 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bghpk"] Oct 13 09:47:28 crc kubenswrapper[4685]: E1013 09:47:28.764203 4685 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1de46b3c-7e5e-42bd-acf3-98498991275f" containerName="collect-profiles" Oct 13 09:47:28 crc kubenswrapper[4685]: I1013 09:47:28.764222 4685 state_mem.go:107] "Deleted CPUSet assignment" podUID="1de46b3c-7e5e-42bd-acf3-98498991275f" containerName="collect-profiles" Oct 13 09:47:28 crc kubenswrapper[4685]: I1013 09:47:28.764461 4685 memory_manager.go:354] "RemoveStaleState removing state" podUID="1de46b3c-7e5e-42bd-acf3-98498991275f" containerName="collect-profiles" Oct 13 09:47:28 crc kubenswrapper[4685]: I1013 09:47:28.766502 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bghpk" Oct 13 09:47:28 crc kubenswrapper[4685]: I1013 09:47:28.782246 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bghpk"] Oct 13 09:47:28 crc kubenswrapper[4685]: I1013 09:47:28.785350 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25374c2d-aad0-47aa-91d0-343d92f81441-utilities\") pod \"redhat-operators-bghpk\" (UID: \"25374c2d-aad0-47aa-91d0-343d92f81441\") " pod="openshift-marketplace/redhat-operators-bghpk" Oct 13 09:47:28 crc kubenswrapper[4685]: I1013 09:47:28.785442 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25374c2d-aad0-47aa-91d0-343d92f81441-catalog-content\") pod \"redhat-operators-bghpk\" (UID: \"25374c2d-aad0-47aa-91d0-343d92f81441\") " pod="openshift-marketplace/redhat-operators-bghpk" Oct 13 09:47:28 crc kubenswrapper[4685]: I1013 09:47:28.785508 4685 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpzhw\" (UniqueName: \"kubernetes.io/projected/25374c2d-aad0-47aa-91d0-343d92f81441-kube-api-access-jpzhw\") pod \"redhat-operators-bghpk\" (UID: \"25374c2d-aad0-47aa-91d0-343d92f81441\") " pod="openshift-marketplace/redhat-operators-bghpk" Oct 13 09:47:28 crc kubenswrapper[4685]: I1013 09:47:28.887499 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25374c2d-aad0-47aa-91d0-343d92f81441-catalog-content\") pod \"redhat-operators-bghpk\" (UID: \"25374c2d-aad0-47aa-91d0-343d92f81441\") " pod="openshift-marketplace/redhat-operators-bghpk" Oct 13 09:47:28 crc kubenswrapper[4685]: I1013 09:47:28.887628 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpzhw\" (UniqueName: \"kubernetes.io/projected/25374c2d-aad0-47aa-91d0-343d92f81441-kube-api-access-jpzhw\") pod \"redhat-operators-bghpk\" (UID: \"25374c2d-aad0-47aa-91d0-343d92f81441\") " pod="openshift-marketplace/redhat-operators-bghpk" Oct 13 09:47:28 crc kubenswrapper[4685]: I1013 09:47:28.887719 4685 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25374c2d-aad0-47aa-91d0-343d92f81441-utilities\") pod \"redhat-operators-bghpk\" (UID: \"25374c2d-aad0-47aa-91d0-343d92f81441\") " pod="openshift-marketplace/redhat-operators-bghpk" Oct 13 09:47:28 crc kubenswrapper[4685]: I1013 09:47:28.888009 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25374c2d-aad0-47aa-91d0-343d92f81441-catalog-content\") pod \"redhat-operators-bghpk\" (UID: \"25374c2d-aad0-47aa-91d0-343d92f81441\") " pod="openshift-marketplace/redhat-operators-bghpk" Oct 13 09:47:28 crc kubenswrapper[4685]: I1013 09:47:28.888050 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25374c2d-aad0-47aa-91d0-343d92f81441-utilities\") pod \"redhat-operators-bghpk\" (UID: \"25374c2d-aad0-47aa-91d0-343d92f81441\") " pod="openshift-marketplace/redhat-operators-bghpk" Oct 13 09:47:28 crc kubenswrapper[4685]: I1013 09:47:28.913631 4685 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpzhw\" (UniqueName: \"kubernetes.io/projected/25374c2d-aad0-47aa-91d0-343d92f81441-kube-api-access-jpzhw\") pod \"redhat-operators-bghpk\" (UID: \"25374c2d-aad0-47aa-91d0-343d92f81441\") " pod="openshift-marketplace/redhat-operators-bghpk" Oct 13 09:47:29 crc kubenswrapper[4685]: I1013 09:47:29.090678 4685 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bghpk" Oct 13 09:47:29 crc kubenswrapper[4685]: I1013 09:47:29.658360 4685 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bghpk"] Oct 13 09:47:30 crc kubenswrapper[4685]: I1013 09:47:30.419215 4685 generic.go:334] "Generic (PLEG): container finished" podID="25374c2d-aad0-47aa-91d0-343d92f81441" containerID="261429bf8b7f4c2c7c9371625a22986a5e0f26107f29bc229cc2e26e32afb736" exitCode=0 Oct 13 09:47:30 crc kubenswrapper[4685]: I1013 09:47:30.419369 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bghpk" event={"ID":"25374c2d-aad0-47aa-91d0-343d92f81441","Type":"ContainerDied","Data":"261429bf8b7f4c2c7c9371625a22986a5e0f26107f29bc229cc2e26e32afb736"} Oct 13 09:47:30 crc kubenswrapper[4685]: I1013 09:47:30.419507 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bghpk" event={"ID":"25374c2d-aad0-47aa-91d0-343d92f81441","Type":"ContainerStarted","Data":"bb6392de3b25a4fc67bd7302f5085d8b2aa4ccd98c981a71e66c8375b3151604"} Oct 13 09:47:30 crc kubenswrapper[4685]: I1013 09:47:30.421787 4685 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 13 09:47:31 crc kubenswrapper[4685]: I1013 09:47:31.429754 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bghpk" event={"ID":"25374c2d-aad0-47aa-91d0-343d92f81441","Type":"ContainerStarted","Data":"2e03f88782adb44bb0313c09e714d85e1701c3e83423de511052bd6510b178bb"} Oct 13 09:47:34 crc kubenswrapper[4685]: I1013 09:47:34.381764 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mzp24/must-gather-d64pf"] Oct 13 09:47:34 crc kubenswrapper[4685]: I1013 09:47:34.382516 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-mzp24/must-gather-d64pf" podUID="5fc0e95d-4f82-4d0f-a745-001486489f27" containerName="copy" containerID="cri-o://cf3fea82d1a3b59f247721e3b5269489c7096919053ec77f06e90300ab2f13f4" gracePeriod=2 Oct 13 09:47:34 crc kubenswrapper[4685]: I1013 09:47:34.393903 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mzp24/must-gather-d64pf"] Oct 13 09:47:34 crc kubenswrapper[4685]: I1013 09:47:34.503686 4685 scope.go:117] "RemoveContainer" containerID="d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790" Oct 13 09:47:34 crc kubenswrapper[4685]: E1013 09:47:34.504108 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:47:34 crc kubenswrapper[4685]: I1013 09:47:34.873497 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mzp24_must-gather-d64pf_5fc0e95d-4f82-4d0f-a745-001486489f27/copy/0.log" Oct 13 09:47:34 crc kubenswrapper[4685]: I1013 09:47:34.874192 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mzp24/must-gather-d64pf" Oct 13 09:47:34 crc kubenswrapper[4685]: I1013 09:47:34.978956 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5fc0e95d-4f82-4d0f-a745-001486489f27-must-gather-output\") pod \"5fc0e95d-4f82-4d0f-a745-001486489f27\" (UID: \"5fc0e95d-4f82-4d0f-a745-001486489f27\") " Oct 13 09:47:34 crc kubenswrapper[4685]: I1013 09:47:34.979050 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lbcz\" (UniqueName: \"kubernetes.io/projected/5fc0e95d-4f82-4d0f-a745-001486489f27-kube-api-access-9lbcz\") pod \"5fc0e95d-4f82-4d0f-a745-001486489f27\" (UID: \"5fc0e95d-4f82-4d0f-a745-001486489f27\") " Oct 13 09:47:34 crc kubenswrapper[4685]: I1013 09:47:34.986351 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fc0e95d-4f82-4d0f-a745-001486489f27-kube-api-access-9lbcz" (OuterVolumeSpecName: "kube-api-access-9lbcz") pod "5fc0e95d-4f82-4d0f-a745-001486489f27" (UID: "5fc0e95d-4f82-4d0f-a745-001486489f27"). InnerVolumeSpecName "kube-api-access-9lbcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:47:35 crc kubenswrapper[4685]: I1013 09:47:35.081375 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lbcz\" (UniqueName: \"kubernetes.io/projected/5fc0e95d-4f82-4d0f-a745-001486489f27-kube-api-access-9lbcz\") on node \"crc\" DevicePath \"\"" Oct 13 09:47:35 crc kubenswrapper[4685]: I1013 09:47:35.089961 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fc0e95d-4f82-4d0f-a745-001486489f27-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "5fc0e95d-4f82-4d0f-a745-001486489f27" (UID: "5fc0e95d-4f82-4d0f-a745-001486489f27"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:47:35 crc kubenswrapper[4685]: I1013 09:47:35.183042 4685 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5fc0e95d-4f82-4d0f-a745-001486489f27-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 13 09:47:35 crc kubenswrapper[4685]: I1013 09:47:35.556615 4685 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mzp24_must-gather-d64pf_5fc0e95d-4f82-4d0f-a745-001486489f27/copy/0.log" Oct 13 09:47:35 crc kubenswrapper[4685]: I1013 09:47:35.560699 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fc0e95d-4f82-4d0f-a745-001486489f27" path="/var/lib/kubelet/pods/5fc0e95d-4f82-4d0f-a745-001486489f27/volumes" Oct 13 09:47:35 crc kubenswrapper[4685]: I1013 09:47:35.560865 4685 generic.go:334] "Generic (PLEG): container finished" podID="5fc0e95d-4f82-4d0f-a745-001486489f27" containerID="cf3fea82d1a3b59f247721e3b5269489c7096919053ec77f06e90300ab2f13f4" exitCode=143 Oct 13 09:47:35 crc kubenswrapper[4685]: I1013 09:47:35.560983 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mzp24/must-gather-d64pf" Oct 13 09:47:35 crc kubenswrapper[4685]: I1013 09:47:35.562021 4685 scope.go:117] "RemoveContainer" containerID="cf3fea82d1a3b59f247721e3b5269489c7096919053ec77f06e90300ab2f13f4" Oct 13 09:47:35 crc kubenswrapper[4685]: I1013 09:47:35.571728 4685 generic.go:334] "Generic (PLEG): container finished" podID="25374c2d-aad0-47aa-91d0-343d92f81441" containerID="2e03f88782adb44bb0313c09e714d85e1701c3e83423de511052bd6510b178bb" exitCode=0 Oct 13 09:47:35 crc kubenswrapper[4685]: I1013 09:47:35.571774 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bghpk" event={"ID":"25374c2d-aad0-47aa-91d0-343d92f81441","Type":"ContainerDied","Data":"2e03f88782adb44bb0313c09e714d85e1701c3e83423de511052bd6510b178bb"} Oct 13 09:47:35 crc kubenswrapper[4685]: I1013 09:47:35.619175 4685 scope.go:117] "RemoveContainer" containerID="de9ab10066e4fb70062591e9cee78610af292f02240bc5a8f175ce5d92ab3aa3" Oct 13 09:47:35 crc kubenswrapper[4685]: I1013 09:47:35.657618 4685 scope.go:117] "RemoveContainer" containerID="cf3fea82d1a3b59f247721e3b5269489c7096919053ec77f06e90300ab2f13f4" Oct 13 09:47:35 crc kubenswrapper[4685]: E1013 09:47:35.658084 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf3fea82d1a3b59f247721e3b5269489c7096919053ec77f06e90300ab2f13f4\": container with ID starting with cf3fea82d1a3b59f247721e3b5269489c7096919053ec77f06e90300ab2f13f4 not found: ID does not exist" containerID="cf3fea82d1a3b59f247721e3b5269489c7096919053ec77f06e90300ab2f13f4" Oct 13 09:47:35 crc kubenswrapper[4685]: I1013 09:47:35.658118 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf3fea82d1a3b59f247721e3b5269489c7096919053ec77f06e90300ab2f13f4"} err="failed to get container status \"cf3fea82d1a3b59f247721e3b5269489c7096919053ec77f06e90300ab2f13f4\": rpc error: code = NotFound desc = could not find container \"cf3fea82d1a3b59f247721e3b5269489c7096919053ec77f06e90300ab2f13f4\": container with ID starting with cf3fea82d1a3b59f247721e3b5269489c7096919053ec77f06e90300ab2f13f4 not found: ID does not exist" Oct 13 09:47:35 crc kubenswrapper[4685]: I1013 09:47:35.658144 4685 scope.go:117] "RemoveContainer" containerID="de9ab10066e4fb70062591e9cee78610af292f02240bc5a8f175ce5d92ab3aa3" Oct 13 09:47:35 crc kubenswrapper[4685]: E1013 09:47:35.659059 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de9ab10066e4fb70062591e9cee78610af292f02240bc5a8f175ce5d92ab3aa3\": container with ID starting with de9ab10066e4fb70062591e9cee78610af292f02240bc5a8f175ce5d92ab3aa3 not found: ID does not exist" containerID="de9ab10066e4fb70062591e9cee78610af292f02240bc5a8f175ce5d92ab3aa3" Oct 13 09:47:35 crc kubenswrapper[4685]: I1013 09:47:35.659081 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de9ab10066e4fb70062591e9cee78610af292f02240bc5a8f175ce5d92ab3aa3"} err="failed to get container status \"de9ab10066e4fb70062591e9cee78610af292f02240bc5a8f175ce5d92ab3aa3\": rpc error: code = NotFound desc = could not find container \"de9ab10066e4fb70062591e9cee78610af292f02240bc5a8f175ce5d92ab3aa3\": container with ID starting with de9ab10066e4fb70062591e9cee78610af292f02240bc5a8f175ce5d92ab3aa3 not found: ID does not exist" Oct 13 09:47:36 crc kubenswrapper[4685]: I1013 09:47:36.586062 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bghpk" event={"ID":"25374c2d-aad0-47aa-91d0-343d92f81441","Type":"ContainerStarted","Data":"9443c3f950c35c6978dddf1a084ab9e7d7461c4c21023b18c2cd69c6ed49b4be"} Oct 13 09:47:39 crc kubenswrapper[4685]: I1013 09:47:39.090968 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bghpk" Oct 13 09:47:39 crc kubenswrapper[4685]: I1013 09:47:39.091328 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bghpk" Oct 13 09:47:40 crc kubenswrapper[4685]: I1013 09:47:40.137320 4685 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bghpk" podUID="25374c2d-aad0-47aa-91d0-343d92f81441" containerName="registry-server" probeResult="failure" output=< Oct 13 09:47:40 crc kubenswrapper[4685]: timeout: failed to connect service ":50051" within 1s Oct 13 09:47:40 crc kubenswrapper[4685]: > Oct 13 09:47:45 crc kubenswrapper[4685]: I1013 09:47:45.514626 4685 scope.go:117] "RemoveContainer" containerID="d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790" Oct 13 09:47:45 crc kubenswrapper[4685]: E1013 09:47:45.517140 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:47:49 crc kubenswrapper[4685]: I1013 09:47:49.200953 4685 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bghpk" Oct 13 09:47:49 crc kubenswrapper[4685]: I1013 09:47:49.248726 4685 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bghpk" podStartSLOduration=15.598328829 podStartE2EDuration="21.248705293s" podCreationTimestamp="2025-10-13 09:47:28 +0000 UTC" firstStartedPulling="2025-10-13 09:47:30.421528407 +0000 UTC m=+3775.569404168" lastFinishedPulling="2025-10-13 09:47:36.071904871 +0000 UTC m=+3781.219780632" observedRunningTime="2025-10-13 09:47:36.618248312 +0000 UTC m=+3781.766124083" watchObservedRunningTime="2025-10-13 09:47:49.248705293 +0000 UTC m=+3794.396581064" Oct 13 09:47:49 crc kubenswrapper[4685]: I1013 09:47:49.287966 4685 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bghpk" Oct 13 09:47:49 crc kubenswrapper[4685]: I1013 09:47:49.471196 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bghpk"] Oct 13 09:47:50 crc kubenswrapper[4685]: I1013 09:47:50.730839 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bghpk" podUID="25374c2d-aad0-47aa-91d0-343d92f81441" containerName="registry-server" containerID="cri-o://9443c3f950c35c6978dddf1a084ab9e7d7461c4c21023b18c2cd69c6ed49b4be" gracePeriod=2 Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.215518 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bghpk" Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.335476 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpzhw\" (UniqueName: \"kubernetes.io/projected/25374c2d-aad0-47aa-91d0-343d92f81441-kube-api-access-jpzhw\") pod \"25374c2d-aad0-47aa-91d0-343d92f81441\" (UID: \"25374c2d-aad0-47aa-91d0-343d92f81441\") " Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.335551 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25374c2d-aad0-47aa-91d0-343d92f81441-utilities\") pod \"25374c2d-aad0-47aa-91d0-343d92f81441\" (UID: \"25374c2d-aad0-47aa-91d0-343d92f81441\") " Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.335593 4685 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25374c2d-aad0-47aa-91d0-343d92f81441-catalog-content\") pod \"25374c2d-aad0-47aa-91d0-343d92f81441\" (UID: \"25374c2d-aad0-47aa-91d0-343d92f81441\") " Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.339088 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25374c2d-aad0-47aa-91d0-343d92f81441-utilities" (OuterVolumeSpecName: "utilities") pod "25374c2d-aad0-47aa-91d0-343d92f81441" (UID: "25374c2d-aad0-47aa-91d0-343d92f81441"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.348162 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25374c2d-aad0-47aa-91d0-343d92f81441-kube-api-access-jpzhw" (OuterVolumeSpecName: "kube-api-access-jpzhw") pod "25374c2d-aad0-47aa-91d0-343d92f81441" (UID: "25374c2d-aad0-47aa-91d0-343d92f81441"). InnerVolumeSpecName "kube-api-access-jpzhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.424553 4685 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25374c2d-aad0-47aa-91d0-343d92f81441-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25374c2d-aad0-47aa-91d0-343d92f81441" (UID: "25374c2d-aad0-47aa-91d0-343d92f81441"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.438001 4685 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpzhw\" (UniqueName: \"kubernetes.io/projected/25374c2d-aad0-47aa-91d0-343d92f81441-kube-api-access-jpzhw\") on node \"crc\" DevicePath \"\"" Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.438035 4685 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25374c2d-aad0-47aa-91d0-343d92f81441-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.438044 4685 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25374c2d-aad0-47aa-91d0-343d92f81441-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.746169 4685 generic.go:334] "Generic (PLEG): container finished" podID="25374c2d-aad0-47aa-91d0-343d92f81441" containerID="9443c3f950c35c6978dddf1a084ab9e7d7461c4c21023b18c2cd69c6ed49b4be" exitCode=0 Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.746253 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bghpk" event={"ID":"25374c2d-aad0-47aa-91d0-343d92f81441","Type":"ContainerDied","Data":"9443c3f950c35c6978dddf1a084ab9e7d7461c4c21023b18c2cd69c6ed49b4be"} Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.746291 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bghpk" event={"ID":"25374c2d-aad0-47aa-91d0-343d92f81441","Type":"ContainerDied","Data":"bb6392de3b25a4fc67bd7302f5085d8b2aa4ccd98c981a71e66c8375b3151604"} Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.746323 4685 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bghpk" Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.746337 4685 scope.go:117] "RemoveContainer" containerID="9443c3f950c35c6978dddf1a084ab9e7d7461c4c21023b18c2cd69c6ed49b4be" Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.783628 4685 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bghpk"] Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.790881 4685 scope.go:117] "RemoveContainer" containerID="2e03f88782adb44bb0313c09e714d85e1701c3e83423de511052bd6510b178bb" Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.793266 4685 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bghpk"] Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.816428 4685 scope.go:117] "RemoveContainer" containerID="261429bf8b7f4c2c7c9371625a22986a5e0f26107f29bc229cc2e26e32afb736" Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.877900 4685 scope.go:117] "RemoveContainer" containerID="9443c3f950c35c6978dddf1a084ab9e7d7461c4c21023b18c2cd69c6ed49b4be" Oct 13 09:47:51 crc kubenswrapper[4685]: E1013 09:47:51.878557 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9443c3f950c35c6978dddf1a084ab9e7d7461c4c21023b18c2cd69c6ed49b4be\": container with ID starting with 9443c3f950c35c6978dddf1a084ab9e7d7461c4c21023b18c2cd69c6ed49b4be not found: ID does not exist" containerID="9443c3f950c35c6978dddf1a084ab9e7d7461c4c21023b18c2cd69c6ed49b4be" Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.878619 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9443c3f950c35c6978dddf1a084ab9e7d7461c4c21023b18c2cd69c6ed49b4be"} err="failed to get container status \"9443c3f950c35c6978dddf1a084ab9e7d7461c4c21023b18c2cd69c6ed49b4be\": rpc error: code = NotFound desc = could not find container \"9443c3f950c35c6978dddf1a084ab9e7d7461c4c21023b18c2cd69c6ed49b4be\": container with ID starting with 9443c3f950c35c6978dddf1a084ab9e7d7461c4c21023b18c2cd69c6ed49b4be not found: ID does not exist" Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.878657 4685 scope.go:117] "RemoveContainer" containerID="2e03f88782adb44bb0313c09e714d85e1701c3e83423de511052bd6510b178bb" Oct 13 09:47:51 crc kubenswrapper[4685]: E1013 09:47:51.879186 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e03f88782adb44bb0313c09e714d85e1701c3e83423de511052bd6510b178bb\": container with ID starting with 2e03f88782adb44bb0313c09e714d85e1701c3e83423de511052bd6510b178bb not found: ID does not exist" containerID="2e03f88782adb44bb0313c09e714d85e1701c3e83423de511052bd6510b178bb" Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.879216 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e03f88782adb44bb0313c09e714d85e1701c3e83423de511052bd6510b178bb"} err="failed to get container status \"2e03f88782adb44bb0313c09e714d85e1701c3e83423de511052bd6510b178bb\": rpc error: code = NotFound desc = could not find container \"2e03f88782adb44bb0313c09e714d85e1701c3e83423de511052bd6510b178bb\": container with ID starting with 2e03f88782adb44bb0313c09e714d85e1701c3e83423de511052bd6510b178bb not found: ID does not exist" Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.879239 4685 scope.go:117] "RemoveContainer" containerID="261429bf8b7f4c2c7c9371625a22986a5e0f26107f29bc229cc2e26e32afb736" Oct 13 09:47:51 crc kubenswrapper[4685]: E1013 09:47:51.879608 4685 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"261429bf8b7f4c2c7c9371625a22986a5e0f26107f29bc229cc2e26e32afb736\": container with ID starting with 261429bf8b7f4c2c7c9371625a22986a5e0f26107f29bc229cc2e26e32afb736 not found: ID does not exist" containerID="261429bf8b7f4c2c7c9371625a22986a5e0f26107f29bc229cc2e26e32afb736" Oct 13 09:47:51 crc kubenswrapper[4685]: I1013 09:47:51.879675 4685 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"261429bf8b7f4c2c7c9371625a22986a5e0f26107f29bc229cc2e26e32afb736"} err="failed to get container status \"261429bf8b7f4c2c7c9371625a22986a5e0f26107f29bc229cc2e26e32afb736\": rpc error: code = NotFound desc = could not find container \"261429bf8b7f4c2c7c9371625a22986a5e0f26107f29bc229cc2e26e32afb736\": container with ID starting with 261429bf8b7f4c2c7c9371625a22986a5e0f26107f29bc229cc2e26e32afb736 not found: ID does not exist" Oct 13 09:47:52 crc kubenswrapper[4685]: I1013 09:47:52.980475 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:47:52 crc kubenswrapper[4685]: I1013 09:47:52.980534 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:47:53 crc kubenswrapper[4685]: I1013 09:47:53.521538 4685 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25374c2d-aad0-47aa-91d0-343d92f81441" path="/var/lib/kubelet/pods/25374c2d-aad0-47aa-91d0-343d92f81441/volumes" Oct 13 09:47:57 crc kubenswrapper[4685]: I1013 09:47:57.503788 4685 scope.go:117] "RemoveContainer" containerID="d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790" Oct 13 09:47:57 crc kubenswrapper[4685]: E1013 09:47:57.505058 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:48:12 crc kubenswrapper[4685]: I1013 09:48:12.503859 4685 scope.go:117] "RemoveContainer" containerID="d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790" Oct 13 09:48:12 crc kubenswrapper[4685]: E1013 09:48:12.504616 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:48:22 crc kubenswrapper[4685]: I1013 09:48:22.980402 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:48:22 crc kubenswrapper[4685]: I1013 09:48:22.981083 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:48:25 crc kubenswrapper[4685]: I1013 09:48:25.515949 4685 scope.go:117] "RemoveContainer" containerID="d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790" Oct 13 09:48:25 crc kubenswrapper[4685]: E1013 09:48:25.516676 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:48:38 crc kubenswrapper[4685]: I1013 09:48:38.503064 4685 scope.go:117] "RemoveContainer" containerID="d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790" Oct 13 09:48:38 crc kubenswrapper[4685]: E1013 09:48:38.507764 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:48:52 crc kubenswrapper[4685]: I1013 09:48:52.980797 4685 patch_prober.go:28] interesting pod/machine-config-daemon-xrvp5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 09:48:52 crc kubenswrapper[4685]: I1013 09:48:52.981307 4685 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 09:48:52 crc kubenswrapper[4685]: I1013 09:48:52.981353 4685 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" Oct 13 09:48:52 crc kubenswrapper[4685]: I1013 09:48:52.982061 4685 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"78e305b74ad025a53109325e9b144a07a1448d8fcb50659d2a199e0a7538b084"} pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 09:48:52 crc kubenswrapper[4685]: I1013 09:48:52.982115 4685 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerName="machine-config-daemon" containerID="cri-o://78e305b74ad025a53109325e9b144a07a1448d8fcb50659d2a199e0a7538b084" gracePeriod=600 Oct 13 09:48:53 crc kubenswrapper[4685]: E1013 09:48:53.109503 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:48:53 crc kubenswrapper[4685]: I1013 09:48:53.411394 4685 generic.go:334] "Generic (PLEG): container finished" podID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" containerID="78e305b74ad025a53109325e9b144a07a1448d8fcb50659d2a199e0a7538b084" exitCode=0 Oct 13 09:48:53 crc kubenswrapper[4685]: I1013 09:48:53.411439 4685 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" event={"ID":"505637be-a3fb-4b68-bd17-9f0ed875fb3c","Type":"ContainerDied","Data":"78e305b74ad025a53109325e9b144a07a1448d8fcb50659d2a199e0a7538b084"} Oct 13 09:48:53 crc kubenswrapper[4685]: I1013 09:48:53.411470 4685 scope.go:117] "RemoveContainer" containerID="a75682cfcb35d15ac8d83500d0d93801124ebd38b4b8e7001d309f70d4ce10ad" Oct 13 09:48:53 crc kubenswrapper[4685]: I1013 09:48:53.412280 4685 scope.go:117] "RemoveContainer" containerID="78e305b74ad025a53109325e9b144a07a1448d8fcb50659d2a199e0a7538b084" Oct 13 09:48:53 crc kubenswrapper[4685]: E1013 09:48:53.412511 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:48:53 crc kubenswrapper[4685]: I1013 09:48:53.504475 4685 scope.go:117] "RemoveContainer" containerID="d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790" Oct 13 09:48:53 crc kubenswrapper[4685]: E1013 09:48:53.504712 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:49:04 crc kubenswrapper[4685]: I1013 09:49:04.503120 4685 scope.go:117] "RemoveContainer" containerID="78e305b74ad025a53109325e9b144a07a1448d8fcb50659d2a199e0a7538b084" Oct 13 09:49:04 crc kubenswrapper[4685]: E1013 09:49:04.503868 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:49:06 crc kubenswrapper[4685]: I1013 09:49:06.503368 4685 scope.go:117] "RemoveContainer" containerID="d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790" Oct 13 09:49:06 crc kubenswrapper[4685]: E1013 09:49:06.504467 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:49:18 crc kubenswrapper[4685]: I1013 09:49:18.502874 4685 scope.go:117] "RemoveContainer" containerID="78e305b74ad025a53109325e9b144a07a1448d8fcb50659d2a199e0a7538b084" Oct 13 09:49:18 crc kubenswrapper[4685]: E1013 09:49:18.503587 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" Oct 13 09:49:20 crc kubenswrapper[4685]: I1013 09:49:20.503449 4685 scope.go:117] "RemoveContainer" containerID="d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790" Oct 13 09:49:20 crc kubenswrapper[4685]: E1013 09:49:20.504349 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:49:32 crc kubenswrapper[4685]: I1013 09:49:32.502712 4685 scope.go:117] "RemoveContainer" containerID="d33d1b02b5b3a5b58e851719b2d745a280a15e2db429added6d42fa044ffd790" Oct 13 09:49:32 crc kubenswrapper[4685]: E1013 09:49:32.503802 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=manager pod=keystone-operator-controller-manager-f9d897d75-xtnqp_openstack-operators(b001d17a-1aea-44ba-86c5-ba6b312156c1)\"" pod="openstack-operators/keystone-operator-controller-manager-f9d897d75-xtnqp" podUID="b001d17a-1aea-44ba-86c5-ba6b312156c1" Oct 13 09:49:33 crc kubenswrapper[4685]: I1013 09:49:33.503315 4685 scope.go:117] "RemoveContainer" containerID="78e305b74ad025a53109325e9b144a07a1448d8fcb50659d2a199e0a7538b084" Oct 13 09:49:33 crc kubenswrapper[4685]: E1013 09:49:33.504283 4685 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xrvp5_openshift-machine-config-operator(505637be-a3fb-4b68-bd17-9f0ed875fb3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-xrvp5" podUID="505637be-a3fb-4b68-bd17-9f0ed875fb3c" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515073145471024454 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015073145472017372 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015073135377016520 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015073135377015470 5ustar corecore